02-07-2024 17:41 PM - last edited 02-22-2024 09:44 AM
SharePoint Batch Update, Create, & Upsert Template
(Also an option for full dataset synchronizations with the "Full Sync" template below)
Compared to the using the basic Apply to Each, Get items, & Update item approach on large amounts of data, this method requires a small fraction of the API calls towards the daily action limit and drastically reduces flow run-times.
It's currently set to take data from any Excel sheet and update records in SharePoint with corresponding data. It works with any potential Power Automate data-source, including HTTP API GET requests, but Excel provides a simple example.
Part of it works like a Vlookup function where it identifies if a row of updated data in Excel or another datasource matches another SharePoint key column and gets the SharePoint record ID for that match. Then it uses the batch update method to update those SharePoint records and it uses the batch create method to create new items for any records without a match.
David Wyatt's Flow Optimization Post For Loading SharePoint Records: https://www.linkedin.com/pulse/top-5-ways-optimize-your-flows-david-wyatt/?trackingId=X9bMmnTZ2QBuu4...
Microsoft Batch API Documentation: https://learn.microsoft.com/en-us/sharepoint/dev/sp-add-ins/make-batch-requests-with-the-rest-apis
TachyTelic.Net Blog & Videos
SharePoint Batch Create Flow
Blog: https://www.tachytelic.net/2021/06/power-automate-flow-batch-create-sharepoint-list-items/
Video: https://youtu.be/2dV7fI4GUYU
SharePoint Batch Delete Flow
Blog: https://www.tachytelic.net/2021/06/power-automate-delete-sharepoint-items/
Video: https://www.youtube.com/watch?v=2ImkuGpEeoo
Version 2.7 - Upsert
-Includes a batch create segment to create an upsert capability. If anyone wants to only update records, then they can remove the Batch Create section. If anyone wants to only create records, then they can go to the GenerateSPData action, remove the expression for the ID field and insert the null value expression.
-Further simplifies the set-up, removing the need to add any additional SharePoint Get items actions & removing the need for parallel branches.
-Can now work on lists with a few million items without adding more actions or branches. It also implements a faster load method using the SharePoint HTTP action as described in point 5 of this article.
-The batch loops have been changed from Do until loops to chunking into Apply to each loops so the batch actions can now run concurrently for additional speed. If you have many batches of data you want to process faster, you can try increasing the concurrency settings on the Apply to each loops containing the SendBatch actions.
-The "setting" inputs action was moved to the top of the flow to help accommodate the new streamlined set-up.
-A SP HTTP call now automatically fixes some issues with referencing the correct list name.
-Faster list load time.
-If you need to batch create &/or update hyperlink columns, check this post
-Adds another HTTP call to get the site users into an object indexed/reference-able by email addresses & gives an example of how to use that to batch update a person column. Anytime the updated source dataset has a blank or an email value not found in the top 5000 site users, it will replace any person in that item with a null value.
Updated set-up screenshots & instructions in this post: https://powerusers.microsoft.com/t5/Power-Automate-Cookbook/Batch-Update-SharePoint-List/m-p/2225500...
Full Sync V1.1 (Combined Upsert & Batch Deletion Sync)
The SharePoint Batch Full Sync template pulls in all the SP List & Source data to perform a batch upsert. But then it also adds on a Batch Deletion Sync to find & delete all the records in the SP List that are not in the Source dataset.
Now, this is initially set up to pull in all the datasource records for all batch actions in the same Do until Get source data loop. And that piece will be limited by the 100MB maximum message / variable size limits for Power Automate, so this Full Sync version will initially only work with datasources with a 100MB or less total size. But this is really only because I'm trying to keep the flow simpler for the majority of users who likely will not have datasources of many 100s of thousands of records.
If you want to further push out against this 100MB limitation, then you will need to separate out the source get data for the batch upsert section from another source get data for the batch deletion sync section. So for the batch upsert section you can use a set-up like in the main batch upsert template where it loads records with all columns 100,000 at a time (or 100 or 5000 or whatever your source dataset per load limitations are) and runs the batch upsert on each source load before running the Do until loop again to get the next source load (which avoids holding anywhere near 100MB in memory at once because it is performing things one load at a time). Then the batch deletion sync section can use a different source get data set-up similar to the "Do until Get destination list IDs + keys" section of the templates where each loop can pull a load from the source dataset & then use a Select action to select only a few of the columns to pass on to the variable holding everything in memory. Since deletions only require the primary key values, you can set the Select to only get the primary key column from each source data load & pass that onto the "Source data outputs" / variable. A full listing of all the primary key values in your source dataset will be much smaller than all columns for the entire table, so that 100MB limit should then hold a few million records worth of the required primary key data to run the batch deletion sync process.
Self Update (See the 1st comment below the main post for the zip download)
The SharePoint Self Batch Update assumes you just want to perform simple updates using only the existing data in the list and removes all the actions related to comparing two datasets to find updates. This may be much easier to use if you just want to quickly do something simple like get all the items created in the past month and mark them all with a Complete status.
But you will be limited to using just the data already in the list and any values you can manually input into the flow.
Version 1.5 - Update
This version makes it easier to handle cases where the list name may have changed since its creation and moves a few of the primary key column matching inputs to the 'settings' compose action so users don't have to look through & edit the more complicated expressions to set up the flow.
The flow can easily expand to any size of SharePoint list by adding more Get items actions and batch update parallel branches. If speed is a concern for anyone, there are ways to make the Get items actions all load in parallel too (up to 50 branches). It's really only limited by how much effort people want to put into their flows & lists.
Google Drive Link to Flow Zip Files: https://drive.google.com/file/d/10p7EB730xsEj-azVYuTIuu8dS0w-AflR/view?usp=sharing
Google Drive Link to Text File to a Scope Action Containing The Flow: https://drive.google.com/file/d/1BVGoeM5mykYlMobAyFkhuLRh3r7jMSLz/view?usp=sharing
Version 1 - Update
Version 1 Explanation Video: https://youtu.be/l0NuYtXdcrQ
Download The Template Batch Update Flow
Google Drive Link to Flow Zip Files: https://drive.google.com/file/d/10gFkycdx6zpRfrI-s_jCDwIK6dpyyDqk/view?usp=sharing
Google Drive Link to Text File to a Scope Action Containing The Flow: https://drive.google.com/file/d/1e6-INUykIT22ppVh5m4kxz8us_7qXy7q/view?usp=sharing
Formulas For Random Number Columns
SharePoint Rand1To50 Column Calculated Default Value Formula:
=INT(INT(RIGHT(NOW(),2))/2)
Excel Random Column Formula:
=ROUNDDOWN(((Rand()*100)+1)/2, 0)
If you have trouble importing any of the flows using the standard legacy import, you can also try importing a Power Apps Solutions package here: Re: Batch Update, Create, and Upsert SharePoint Li... - Page 25 - Power Platform Community (microsof...
Thanks for any feedback,
Please subscribe to my YouTube channel (https://youtube.com/@tylerkolota?si=uEGKko1U8D29CJ86).
And reach out on LinkedIn (https://www.linkedin.com/in/kolota/) if you want to hire me to consult or build more custom Microsoft solutions for you.
watch?v=QCkjQy6sHZg
Check this post & the one following it.
For lookup columns, I think you can only update which lookup item is referenced.
And you set it up similar to the Person column updating the LookupColumnNameId value.
I got the same error. You have to add "Id" at the field name on the left. i.e. if the name of your Sharepoint list lookup field is "Example", you have to populate the left side of GenerateSpData with "ExempleId".
IT was the pagination @takolota , other problem. Trying to update dates I receive this error:
"value": "Cannot convert a primitive value to the expected type 'Edm.DateTime'. See the inner exception for more details."
For what I understand it because the date is in this format "31/01/2023" and should be in this "2024-01-31T00:00:00.000Z\". But I don't know how to put this. Can you help me?
When you expand the advanced options on the Excel List rows, do you have the date format parameter set to ISO?
Even if you do, Excel sometimes still returns serial numbers instead of dates. If that happens you can use the TachyTelic method to convert the serial number to dates.
https://www.tachytelic.net/2020/11/convert-excel-dates-power-automate/?amp
Thanks @takolota , I already fix it. Now appear some errors like this one: "value":
"The type SP.ListItemEntityCollection does not support HTTP PATCH method." or
"Your changes conflict with those made concurrently by another user. If you want your changes to be applied, click Back in your Web browser, refresh the page, and resubmit your changes." or
HTTP/1.1 500 Internal Server Error
CONTENT-TYPE: application/json;odata=verbose;charset=utf-8 or
HTTP/1.1 409 Conflict
CONTENT-TYPE: application/json;odata=verbose;charset=utf-8
What can I do?
Thank you @takolota , you solve the problem. I have a question, could I apply these steps just to export the sharepoint list to an excel file in batch, or do you know any away to do it? Thank you very much
To export SharePoint lists to Excel there are several examples by now…
I have my own batch Excel set up: https://powerusers.microsoft.com/t5/Power-Automate-Cookbook/Excel-Batch-Create-Update-and-Upsert/td-...
And CSV to new Excel table set up: https://powerusers.microsoft.com/t5/Power-Automate-Cookbook/CSV-To-New-Excel-Table/td-p/1826096
@DamoBird365 also recently did a video showing these types of methods: https://youtu.be/gtlklzi6MDg?si=iEpE4V-XqZyQaqfX
Thank you for getting back to me. I have to preface this with I am a complete novice with no training in any of the Microsoft power applications. But I had a requirement to create validations and shopping carts etc. for my work. I came across your video when I was searching for a way to daily upload a file exported in excel to a SharePoint list so it could be used as a Datasource during a validation step in a power apps shopping cart. Again, completely self-taught and extremely thankful to all the real programmers out there that I copy. So I used your flow as you created it with my info and everything worked out great if I went in everyday to reconnect the excel file name and table name in the List items Present in a Table step. My co-works couldn't figure out how to keep that updated every day so for about a week we had old data in the validation process so that kicked off my search how to automate the process and to automatically grab the excel file name and table name. I should first say I am working on a GCC account. So my entire process starts from an email:
Flow 1 grabs the excel file from the email and adds it to a OneDrive folder.
Flow 2 Renames the file (because the name changes every day) and then Creates a table Renames and Moves the file again to another folder in OneDrive
Flow 3 Moves the file to a SharePoint library.
and then I changed you flow to look like this, and it works now. When I wrote you, I was struggling to make the flow function.
Someone could more than likely write it better but at least now it doesn't matter if the file is overwritten in the SharePoint library the file is overwritten with the same file name and table name everyday and it dynamically captures the file name and table name and writes back to the SharePoint list.
Hi @hstep (and @takolota ). Could you explain exactly how you sorted out your issue when switching to a scheduled flow? I'm seeing the same problem (using the v2.7 flow). It all works fine from a manual trigger, but when I switch to a scheduled trigger the flow complains about the limits on the get rows action and won't run without that being fixed. If I keep the manual trigger and switch the get rows 'top count', 'skip count', 'pagination threshold' to 5000, as well as 'do until upsert' length comparison to be less than 5000, then that still works (I have 5500 records at the moment). But doing the same with a manual trigger doesn't work. It clears the error messages, but just leads to the test run seeming to hang on the 'Do until Upsert' action (and, frustratingly, if I cancel then no output is generated from that step to help narrow down the problem). Is there something else I should be tweaking?