Click here to Skip to main content
15,115,667 members
Articles / Artificial Intelligence / Machine Learning
Posted 25 Nov 2021

Tagged as


1 bookmarked

How to achieve Landmark Recognition With Machine Learning

Rate me:
Please Sign up or sign in to vote.
3.12/5 (3 votes)
25 Nov 2021Apache3 min read

Ever seen a breathtaking landmark or scenery while flipping through a book or magazine, and been frustrated because you don't know what it's called or where it is? Wouldn't it be great if there was an app that could tell you what you're seeing! Fortunately, machine learning makes it remarkably easy to develop such an app. 

Introduction to Landmark Recognition​

The landmark recognition service enables you to obtain the landmark name, landmark longitude and latitude, and even a confidence value of the input image. When you input an image for recognition, a confidence value will be provided whereby a higher confidence value indicates that the landmark in the input image is more likely to be recognized. You can then use this information to create a highly-personalized experience for your users. 

When using landmark recognition, the device calls the on-cloud API for detection, and the detection algorithm model runs on the cloud. You'll need to ensure that the device is connected to the Internet while using this service.


Configuring the development environment

1. Create an app in AppGallery Connect.

Image 1

For details, see Getting Started with Android.

2. Enable ML Kit.

Image 2

Click here for more details.

3. Download the agconnect-services.json file, which is automatically generated after the app is created. Copy it to the app directory of the Android Studio project.

Image 3

Image 4

4. Configure the Maven repository address for the HMS Core SDK.
5. Integrate the landmark recognition SDK.
Configure the SDK in the build.gradle file in the app directory.

<code>// Import the landmark recognition SDK.
implementation 'com.huawei.hms:ml-computer-vision-cloud:'</code>

Add the AppGallery Connect plugin configuration as needed through either of the following methods:
Method 1: Add the following information under the declaration in the file header:

<code>apply plugin: ''
apply plugin: 'com.huawei.agconnect'</code>

Method 2: Add the plugin configuration in the plugins block:

<code>plugins {
    id ''
    id 'com.huawei.agconnect'

Code Development
1. Obtain the camera permission to use the camera.
(Mandatory) Set the static permission.

<code><uses-permission android:name="android.permission.CAMERA" /></code>

(Mandatory) Obtain the permission dynamically.

        this, new String[]{Manifest.permission. CAMERA

}, 1);</code>

2. Set the API key. This service runs on the cloud, meaning an API key is required to set the cloud authentication information for the app. This step is mandatory, and failure to complete it will result in an error being reported when the app is running.

<code>// Set the API key to access the on-cloud services.
private void setApiKey() {

// Parse the agconnect-services.json file to obtain its information.
AGConnectServicesConfig config = AGConnectServicesConfig.fromContext(getApplication());
// Sets the API key.


3. Create a landmark analyzer through either of the following methods:

<code>// Method 1: Use default parameter settings.

MLRemoteLandmarkAnalyzer analyzer = MLAnalyzerFactory.getInstance().getRemoteLandmarkAnalyzer();</code>
<code>// Method 2: Use customized parameter settings through the MLRemoteLandmarkAnalyzerSetting class.


 * Use custom parameter settings.
 * setLargestNumOfReturns indicates the maximum number of recognition results.
 * setPatternType indicates the analyzer mode.
 * MLRemoteLandmarkAnalyzerSetting.STEADY_PATTERN: The value 1 indicates the stable mode.
 * MLRemoteLandmarkAnalyzerSetting.NEWEST_PATTERN: The value 2 indicates the latest mode.
private void initLandMarkAnalyzer() {
    settings = new MLRemoteLandmarkAnalyzerSetting.Factory()
    analyzer = MLAnalyzerFactory.getInstance().getRemoteLandmarkAnalyzer(settings);

4. Convert the image collected from the camera or album to a bitmap. This is not provided by the landmark recognition SDK, so you'll need to implement it on your own.

<code>// Select an image.
private void selectLocalImage() {
    Intent intent = new Intent(Intent.ACTION_PICK, null);
    intent.setDataAndType(MediaStore.Images.Media.EXTERNAL_CONTENT_URI, "image/*");
    startActivityForResult(intent, REQUEST_SELECT_IMAGE);

Enable the landmark recognition service in the callback.

protected void onActivityResult(int requestCode, int resultCode, Intent data) {
    super.onActivityResult(requestCode, resultCode, data);
    // Image selection succeeded.
    if (requestCode == REQUEST_SELECT_IMAGE && resultCode == RESULT_OK) {
        if (data != null) {
            // Obtain the image URI through getData().              imageUri = data.getData();

// Implement the BitmapUtils class by yourself. Obtain the bitmap of the image with its URI.

bitmap = BitmapUtils.loadFromPath(this, imageUri, getMaxWidthOfImage(), getMaxHeightOfImage());
        // Start landmark recognition.

5. Start landmark recognition after obtaining the bitmap of the image. As this service runs on the cloud, a poor network connection may slow down data transmission. Therefore, it's recommended that you add a mask to the bitmap prior to landmark recognition.

<code>// Start landmark recognition.
private void startAnalyzerImg(Bitmap bitmap) {
    if (imageUri == null) {
    // Add a mask.

    // Create an MLFrame object using JPG, JPEG, PNG, and BMP images are supported. It is recommended that the image size be greater than or equal to 640 x 640 px.
    MLFrame mlFrame = new MLFrame.Creator().setBitmap(bitmap).create();
    Task<List<MLRemoteLandmark>> task = analyzer.asyncAnalyseFrame(mlFrame);
    task.addOnSuccessListener(new OnSuccessListener<List<MLRemoteLandmark>>() {
        public void onSuccess(List<MLRemoteLandmark> landmarkResults) {
            // Called upon recognition success.
            Log.d("BitMapUtils", landmarkResults.get(0).getLandmark());
    }).addOnFailureListener(new OnFailureListener() {
        public void onFailure(Exception e) {
            // Called upon recognition failure.
            // Recognition failure.
            try {
                MLException mlException = (MLException) e;
                // Obtain the result code. You can process the result code and set a different prompt for users for each result code.
                int errorCode = mlException.getErrCode();
                // Obtain the error message. You can quickly locate the fault based on the result code.
                String errorMessage = mlException.getMessage();
                // Record the code and message of the error in the log.
                Log.d("BitMapUtils", "errorCode: " + errorCode + "; errorMessage: " + errorMessage);
            } catch (Exception error) {
                // Handle the conversion error.

Testing the App​

The following illustrates how the service works, using the Oriental Pearl Tower in Shanghai and Pyramid of Menkaure as examples:

Image 5

Image 6

More Information​

1. Before performing landmark recognition, set the API key to set the cloud authentication information for the app. Otherwise, an error will be reported while the app is running.
2. Landmark recognition runs on the cloud, so completion may be slow. It is recommended that you add a mask before performing landmark recognition.


This article, along with any associated source code and files, is licensed under The Apache License, Version 2.0


About the Author

Comments and Discussions

GeneralMy vote of 5 Pin
Ștefan-Mihai MOGA25-Nov-21 21:18
professionalȘtefan-Mihai MOGA25-Nov-21 21:18 

General General    News News    Suggestion Suggestion    Question Question    Bug Bug    Answer Answer    Joke Joke    Praise Praise    Rant Rant    Admin Admin   

Use Ctrl+Left/Right to switch messages, Ctrl+Up/Down to switch threads, Ctrl+Shift+Left/Right to switch pages.