Tag: nijos.dev

I Wish I Would Have Known #31: App Access Checker

App access checker image

You can enter a user’s email address and see the list of published apps in your environment and all the access, license, and security information specific to that user. This can be very useful in troubleshooting why a user cannot see a specific app in your environment. The tool can be accessed from the admin center or via direct URL.


Hope this helps..

Early Bound vs Late Bound Programming in Dataverse

image of a computer screen with programming code

As you all know, when working with code for Microsoft Dataverse we have two options late-bound and early-bound. I have seen many people advocating one option over another and explaining why one is better than the other. For me, it’s just a personal choice, if you prefer typed classes go for early bound else go for the late-bound approach. Add some additional details below to help you to select a programming option.

Early Bound

Early-bound programming requires that you first generate a set of classes based on the table and column definitions (entity and attribute schema) for a specific environment using the code generation tool. Once generated you can enjoy better IntelliSense and compile-time checking of all types so that no implicit casts occur. Below is the sample early bind code from Microsoft Docs which creates an account record.

var account = new Account();
// set attribute values
    // string primary name
    account.Name = "Contoso";
    // Boolean (Two option)
    account.CreditOnHold = false;
    // DateTime
    account.LastOnHoldTime = new DateTime(2017, 1, 1);
    // Double
    account.Address1_Latitude = 47.642311;
    account.Address1_Longitude = -122.136841;
    // Int
    account.NumberOfEmployees = 500;
    // Money
    account.Revenue = new Money(new decimal(5000000.00));
    // Picklist (Option set)
    account.AccountCategoryCode = new OptionSetValue(1); //Preferred customer

//Create the account
Guid accountid = svc.Create(account);

Late Bound

Here you don’t need to generate the classes, you directly start wrting your logic, However, you need to refer to tables and columns (entities and attributes) using their LogicalNames in the Entity class. You don’t get compile-time validation on names of entities, attributes, and relationships. Types are checked only when the object is created. Below is the sample late bind code from Microsoft Docs which creates an account record.

//Use Entity class specifying the entity logical name
var account = new Entity("account");

// set attribute values
    // string primary name
    account["name"] = "Contoso";            
    // Boolean (Two option)
    account["creditonhold"] = false;
    // DateTime
    account["lastonholdtime"] = new DateTime(2017, 1, 1);
    // Double
    account["address1_latitude"] = 47.642311;
    account["address1_longitude"] = -122.136841;
    // Int
    account["numberofemployees"] = 500;
    // Money
    account["revenue"] = new Money(new decimal(5000000.00));
    // Picklist (Option set)
    account["accountcategorycode"] = new OptionSetValue(1); //Preferred customer
//Create the account
Guid accountid = svc.Create(account);

Now let’s do a comparison on both.

You can verify entity, attribute, and relationship names at compile timeNo compile-time verification of entity, attribute, and relationship names
You must generate entity classesYou don’t need to generate entity classes
Better IntelliSense supportLess IntelliSense support
Less readable codeMore readable code
Slightly less performantSlightly more performant
Generated classes are quite largeGenerated classes are small

Hope this helps.

Event Execution Pipeline in Dataverse/Dynamics 365

When the standard low-code options cannot handle your business logic in an expected manner, as a developer, you can use custom event handlers in Microsoft Dataverse. The Event Framework provides the capability to register custom code to be run in response to specific events.

The custom code options available are Plugins, Azure integrations, virtual table data providers, and Webhooks, etc… When configured correctly using the plugin registration tool, these custom code extensions will respond to events that are provided by the event framework.

When an event occurs a message is sent to the organization Web service to process the event. The message contains information about the event, such as details of the table where the event occurred, dataverse organization details, details of the user who triggered the event, etc… There is a specific message for each event such as create, update, retrieve, retrieve multiple, associate, disassociate, delete, etc… Each message is processed in a series of 4 stages called the event execution pipeline and your custom code extension can be registered at any of these stages based on your business need. See the below table for the details about each stage.

Stage NameDescription
PreValidationThis stage will occur before the main system operation. This provides an opportunity to include logic to cancel the operation before the database transaction.
This stage occurs before any security checks are performed to verify that the calling or logged-on user has the correct permission to perform the intended operation.
PreOperationOccurs before the main system operation and within the database transaction. If you want to change any values for an entity included in the message, you should do it here.

Avoid canceling an operation here. Canceling will trigger a rollback of the transaction and have a significant performance impact.
MainOperationFor internal use only except for Custom API and Custom virtual table data providers.
More information:
Create and use Custom APIs
Custom Virtual table data providers
PostOperationOccurs after the main system operation and within the database transaction. Use this stage to modify any properties of the message before it is returned to the caller.

Avoid applying changes to an entity included in the message because this will trigger a new Update event.

Within the PostOperation stage, you can register steps to use the asynchronous execution mode. These steps will run outside of the database transaction using the asynchronous service. More information Asynchronous service.

Hope this helps.

Call AI Models Using Power FX

Power Fx is the general-purpose, strong-typed, declarative, and functional programming language used across the Microsoft Power Platform. This excel-like formula language helps app makers to create apps using low code, read more about Power FX here.

With the latest announcement, Microsoft is giving another level of possibilities with Power FX. Now, you can bind the AI model to any control using Power FX. You heard it right! any control. Earlier it was possible to use AI with the AI builder controls only. Now, Microsoft has made it easier than ever to add intelligence to your Power Apps directly through Power FX. Let’s see how to use this feature for a text control to detect the language entered.

Turn on the AI models as data sources option

Got App setting and turn on AI model as date source

Figure1: Turn on Ai model as data source

In your power app, under the app settings go to Upcoming features > Preview, and turn on the AI model as data sources.

Add your AI model as Data Source

Under the Data tab, go to Add data and In the AI models list, select a language detection model. To know more about AI models please check the official Microsoft docs here.

Screenshot of how to select your model.

Fig 2: Add your AI model to the APP

Add your Controls

Place a text input and two text labels on the canvas as explained below

  1. Rename the text input to TextInput1.
  2. Select Text label and place it on the canvas.
  3. Rename the text label to Language.
  4. Add another text label by selecting Text label and place it to the right of the Language text label.

Call the AI model using Power FX

Now we can use the power FX formula to call the AI model method. Select the text label you added in step 2 and enter the following Power Fx formula.

First('Language detection'.Predict(TextInput1.Text).results).language
Screenshot of the Power Fx formula.
Fig 3 : Power FX formaula

Now, save and run the app. You can type in random languages in the text box to see the AI model in action.

Screenshot of how to try out the app you created.
Fig 4: language prediction.

Now you have bound the AI model to a normal label control using Power FX.

Formula explained

‘Language detection’.Predict(TextInput1.Text).results returns a table of possible languages detected with the score. The language with the highest score (highest possible match) will be the first item in the table. So, using the function ‘First‘ the most matching language code is set to the label.

Another example of a prediction model can be found here.

Source: https://docs.microsoft.com/en-us/ai-builder/powerfx-in-powerapps

Hope this helps.

PowerAutomate: Teams Connector Fails to List TimeZones

Connector error saying cannot list timezones.

Recently, I faced a strange issue, I was trying to create teams meeting using cloud flow, and the connector’s create meeting step was not listing timezone and not allowing me to set timezone as a custom dynamic value. It was also giving a retrieval error message as shown below.

Could not retrieve values. The dynamic invocation request failed with error: {
  "error": {
    "code": "ResourceNotFound",
    "message": "Resource could not be discovered.",
    "innerError": {
      "date": "2021-10-28T11:49:20",
      "request-id": "a0e7df5f-d119-4534-a8c3-5332e6f8b906",
      "client-request-id": "a0e7df5f-d119-4534-a8c3-5332e6f8b906"

The Issue

I was testing this using a service account, and this was working fine for my user account. I was sure from the beginning itself that it was going to be some permission issue. However, I had no idea what permission I should look for. Finally, I managed to find that teams and other Microsoft applications use the Outlook Graph APIS to retrieve timezones and languages. I tried calling the APIs using the service account and I could replicate the error response. Now, it’s clear that the user is missing the required permissions to call outlook graph APIs.

The Fix

I am not so proficient with outlook and I raised this concern with our internal outlook team and they fixed this issue. What I could gather from them is that, the service account’s online office mailbox was not correctly enabled , which then they enabled with proper permissions. If you face similar issue, checking the mailbox permissions could save you a few hours 🙂

Hope this helps.

I Wish I Would Have Known #29: Default Environment Limitations

I Wish I Would Have Known strip 29: Default Environment Limitations

A single default environment is automatically created by Power Apps for each tenant and shared by all users in that tenant. Whenever a new user signs up for Power Apps, they’re automatically added to the Maker role of the default environment. The default environment is created in the region closest to the default region of the Azure AD tenant. Read more