id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
1755730242 | Installtion Setup
Dear Team,
When I am planning to execute the setup getting the following error:
ERROR: deployment failed: failing invoking action 'provision', error deploying infrastructure: deploying to subscription:
Deployment Error Details:
InvalidTemplateDeployment: The template deployment 'openai' is not valid according to the validation procedure. The tracking id is '6777928d-1567-4630-b286-8f77cbf7cacf'. See inner errors for details.
InsufficientQuota: The specified capacity '120' of account deployment is bigger than available capacity '0' for UsageName 'Tokens Per Minute (thousands) - GPT-35-Turbo'.
TraceID: 09f9bd8b6cdac42e96b5848442b7a646
But here I am trying to use the existing the deployed gpt & chat model in spite of why it is going to create new deployment model?
AZURE_OPENAI_CHATGPT_DEPLOYMENT="davinci"
AZURE_OPENAI_GPT_DEPLOYMENT="chat"
AZURE_OPENAI_RESOURCE_GROUP="XXXX"
AZURE_OPENAI_SERVICE="XXXX"
am I missing anything can you please help on the same
Thanks,
Bhaskar
I am getting the same error and I was able to successfully deploy last Friday.
The same here.
same
same
I have the same issue since this week indeed.
If you need 120 for example, but it is currently taking 120/120 , it still demands another 120 units to be free, although no extra units would be needed.
Temporary fix for us is to go to quotas and set all current deployments to a very low number such that you have enough free space again. You'd have to do this for every deployment, so I am also waiting for a fix.
I have the same issue since this week indeed. If you need 120 for example and it is currently taking 120/120 , it still demands another 120 units to be free, although no extra units would be needed in practice. Temporary fix for us is to go to quotas and set all current deployments to a very low number such that you have enough free space again. You'd have to do this for every deployment, so I am also waiting for a fix.
I'm having trouble changing the quotas, where is it?
Thanks for advise I had updated the quota in model the deployment is going fine.. but eventually we need to wait for fix form MSFT.
so, temporary solution is to go to Azure AI studio, Deployments, and for each Edit deployment, Advanced, set "Tokens per Minute Rate Limit" from 120 to e.g. 2. Then deploy your accelerator
This is very inconvenient - having to change quotas manually, especially if the infra is deployed frequently. I hope it's gonna get fixed soon.
How do you decrease the quota for the service you are deploying?
Let's presume that you are using 120/120.
You decrease it to 1/120.
Then it tries to deploy another 120 and complains that it is more than the free quota which is 119.
This https://github.com/Azure-Samples/azure-search-openai-demo/commit/f5a7095e44bd45e224c3200874be36e93e4e0579 should help. Can you either get that patch into your current code or start from scratch and try again?
| gharchive/issue | 2023-06-13T22:18:20 | 2025-04-01T06:36:44.508615 | {
"authors": [
"GregLin0420",
"TimoRiikonenIF",
"bhasampa",
"clojosh",
"jjczopek",
"jongio",
"paprocki-r",
"silvio-machado",
"tickx-cegeka"
],
"repo": "Azure-Samples/azure-search-openai-demo",
"url": "https://github.com/Azure-Samples/azure-search-openai-demo/issues/307",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2378483692 | Azure SQL -> Should be SQL Server, there's no Azure SQL used in this sample project.
Instances of Azure SQL mentioned in README and sample title should reflect that this sample uses SQL Server 2022 (Linux). I'm missing where Azure SQL is used. I supposed you'd need to update the Readme to indicate that you could publish the dacpac to Azure SQL if desired. Then document where and how to update the DB connection string in the sample to point to Azure SQL and not localhost.
Thanks for the feedback! This issue was partially fixed on this commit: 06ff3078ffe49284f0d95fdcc66fcd6b20f5e994
The SQL Database project is created for an Azure SQL Database target, thus the database schema is Azure SQL Database compatible. I added a note to explain this situation here.
| gharchive/issue | 2024-06-27T15:34:01 | 2025-04-01T06:36:44.512839 | {
"authors": [
"croblesm",
"seantleonard"
],
"repo": "Azure-Samples/azure-sql-library-app-blazor",
"url": "https://github.com/Azure-Samples/azure-sql-library-app-blazor/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2300299845 | Remove unnecessary semicolon
Purpose
This remove unnecessary semicolon in Swift.
Pull Request Type
What kind of change does this Pull Request introduce?
[ ] Bugfix
[ ] Feature
[x] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Documentation content changes
[ ] Other... Please describe:
/azp run
| gharchive/pull-request | 2024-05-16T12:40:38 | 2025-04-01T06:36:44.514372 | {
"authors": [
"MoonGoon72",
"yulin-li"
],
"repo": "Azure-Samples/cognitive-services-speech-sdk",
"url": "https://github.com/Azure-Samples/cognitive-services-speech-sdk/pull/2378",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1839422552 | HackTogether: Fix feature listing for registered users
Are you accessing the project from the GitHub Codespaces?
[ ] No, I am running the project on my local machine (I will fill in the OS section below)
[x] Yes, I am using GitHub Codespaces.
Note: The project is optimized to run on Codespaces. We welcome contributions to enhance it for cross-platform local development, but it's not something we will prioritize right now. Thank you!
Describe the bug
The bookmark icon is too close to the image.
The Property description is overlapping the bookmark icon.
The short description for the listed property could be removed in order to create a minimalistic design. This description could be assessed when the user wants to know more about the property.
To Reproduce
Steps to reproduce the behavior
Go to home page after login
Expected behavior
I expect the bookmark icon to be properly aligned/placed
The property name/description should not overlap the bookmark icon
A more aesthetic property listing
Screenshots
OS (please complete the following information):
OS:
OS Version:
npm version:
Node.js version:
output of azd version:
Additional context
Having the long description removed from the featured listing will let each card have equal height.
Hi @mzbhewtee Thanks for reporting! We suggest you investigate this issue as related to Angular material design. Thank you for participating. We look forward to your PR!
Hi @anfibiacreativa Thank you, i will investigate the issue.
Suggestion to tackle with issue #242
Removing long description is not ideal recommendation. we can properly customize scss for having responsive cards with appropriate dimensions.
I am also willing to contribute and investigate together @mzbhewtee
Do we have any updates? Thank you!
| gharchive/issue | 2023-08-07T13:06:36 | 2025-04-01T06:36:44.522400 | {
"authors": [
"Tanmai2002",
"anfibiacreativa",
"juliamuiruri4",
"mzbhewtee"
],
"repo": "Azure-Samples/contoso-real-estate",
"url": "https://github.com/Azure-Samples/contoso-real-estate/issues/264",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2179206035 | fix: updated command for package installs for the notification server (ReadMe)
Issue #451
Readme instruction
Current output
Target output
Hi @juliamuiruri4 your change LGTM. Can you change your commit message to follow these guidelines.
Done @manekinekko
Let me know if I should change anything else. Thanks
| gharchive/pull-request | 2024-03-11T13:57:44 | 2025-04-01T06:36:44.525405 | {
"authors": [
"juliamuiruri4",
"manekinekko"
],
"repo": "Azure-Samples/contoso-real-estate",
"url": "https://github.com/Azure-Samples/contoso-real-estate/pull/452",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
887863890 | Adding pop tokens sample
Purpose
Adds a new sample where React SPA acquiring PoP tokens to call Node.js web API which validates the token using jose
cc @hectormmg
Does this introduce a breaking change?
[ ] Yes
[ x ] No
Pull Request Type
What kind of change does this Pull Request introduce?
[ ] Bugfix
[ ] Feature
[ ] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Documentation content changes
[ x ] Other... Please describe: **new sample**
How to Test
Get the code
git clone https://github.com/Azure-Samples/ms-identity-javascript-react-tutorial.git
cd ms-identity-javascript-react-tutorial
git checkout pop-sample
cd 6-AdvancedScenarios/2-call-api-pop/API
npm install
cd ../SPA
npm install
What to Check
Verify that the following are valid
Test in Chrome, Edge, Firefox
Test in private mode
"dependencies": {
delete this file?
Refers to: 3-Authorization-II/1-call-api/SPA/package-lock.json:6 in 3e8d6a7. [](commit_id = 3e8d6a7d8c3a8f054648fc291b1fd33a202924dc, deletion_comment = False)
| gharchive/pull-request | 2021-05-11T16:50:40 | 2025-04-01T06:36:44.529326 | {
"authors": [
"derisen",
"kalyankrishna1"
],
"repo": "Azure-Samples/ms-identity-javascript-react-tutorial",
"url": "https://github.com/Azure-Samples/ms-identity-javascript-react-tutorial/pull/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1482257365 | Deployment fails when using Custom Networking and when using custom Address space
Describe the bug
When using AKS Construction helper, the deployment fails with the following error:
{"status":"Failed","error":{"code":"DeploymentFailed","message":"At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/DeployOperations for usage details.","details":[{"code":"Conflict","message":"{\r\n \"status\": \"Failed\",\r\n \"error\": {\r\n \"code\": \"ResourceDeploymentFailure\",\r\n \"message\": \"The resource operation completed with terminal provisioning state 'Failed'.\",\r\n \"details\": [\r\n {\r\n \"code\": \"DeploymentFailed\",\r\n \"message\": \"At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/DeployOperations for usage details.\",\r\n \"details\": [\r\n {\r\n \"code\": \"BadRequest\",\r\n \"message\": \"{\\r\\n \\\"error\\\": {\\r\\n \\\"code\\\": \\\"NetcfgInvalidSubnet\\\",\\r\\n \\\"message\\\": \\\"Subnet 'AzureFirewallManagementSubnet' is not valid in virtual network 'vnet-aks-rv-test'.\\\",\\r\\n \\\"details\\\": []\\r\\n }\\r\\n}\"\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n}"}]}}
It appears that even when no Azure Firewall is requested in the helper, the helper attempts to create an AzureFirewallManagementSubnet
To Reproduce
Steps to reproduce the behavior:
Go to Networking Details tab
Under Network plugin, select "Use CNI..."
Under Default or Custom VNET, select Custom Networking
VNET Address space, anything, but the default space. EX: 10.200.0.0/16
AKS Nodes Subnet, anything within the range. EX: 10.200.0.0/20
Alternatively, it can be recreated it by: URL for AKS-Construction helper config
# Create Resource Group
az group create -l EastUS2 -n aks-rv-test-rg
# Deploy template with in-line parameters
az deployment group create -g aks-rv-test-rg --template-uri https://github.com/Azure/AKS-Construction/releases/download/0.9.5/main.json --parameters \
resourceName=aks-rv-test \
kubernetesVersion=1.25.2 \
upgradeChannel=stable \
SystemPoolType=Standard \
nodePoolName=npsymuser01 \
agentCountMax=20 \
osDiskType=Managed \
custom_vnet=true \
vnetAddressPrefix=10.200.0.0/19 \
vnetAksSubnetAddressPrefix=10.200.0.0/21 \
enable_aad=true \
AksDisableLocalAccounts=true \
enableAzureRBAC=true \
adminPrincipalId=$(az ad signed-in-user show --query id --out tsv) \
registries_sku=Standard \
acrPushRolePrincipalId=$(az ad signed-in-user show --query id --out tsv) \
enableTelemetry=false \
omsagent=true \
retentionInDays=30 \
fileCSIDriver=false \
diskCSIDriver=false \
keyVaultAksCSI=true \
keyVaultCreate=true \
keyVaultOfficerRolePrincipalId=$(az ad signed-in-user show --query id --out tsv) \
keyVaultAksCSIPollInterval=5m \
fluxGitOpsAddon=true
Expected behavior
Deployment succeeds, and provisioned VNet does not have an AzureFirewallManagementSubnet created without requesting it.
I think this could have been fixed in #480 and just needs released.
@MattLeach25 - thoughts?
I think this could have been fixed in #480 and just needs released. @MattLeach25 - thoughts?
Yeah this is fixed with the new release.
We'll aim to cut a release tomorrow.
@romerve - if you clone the repo and target bicep/main.bicep instead of the release - then you should be good.
@MattLeach25 created a release today, so this should be fixed.
| gharchive/issue | 2022-12-07T15:58:06 | 2025-04-01T06:36:44.536725 | {
"authors": [
"Gordonby",
"MattLeach25",
"romerve"
],
"repo": "Azure/AKS-Construction",
"url": "https://github.com/Azure/AKS-Construction/issues/482",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1262848073 | Setting codespaces memory requirement
PR Summary
Closes #289
PR Checklist
[x] PR has a meaningful title
[x] Summarized changes
[x] This PR is ready to merge and is not Work in Progress
[x] Link to a filed issue
approve, but no way react dev takes 8GB, only takes 500MB locally
| gharchive/pull-request | 2022-06-07T07:29:06 | 2025-04-01T06:36:44.539255 | {
"authors": [
"Gordonby",
"khowling"
],
"repo": "Azure/AKS-Construction",
"url": "https://github.com/Azure/AKS-Construction/pull/302",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1621778907 | 26808 - Feature - Add support to allow Private DNS Zone VNet Linking for Spoke VNets
Overview/Summary
Adding support to allow Private DNS Zone VNet Linking for Spoke VNets through HubPeeredSpoke orchestration module and a new Private Dns Zone Linking module.
AB#26808
This PR fixes/adds/changes/removes
Added feature to HubPeeredSpoke orchestration module to allow Private DNS Zone Linking to Spoke VNets.
Added new module for Private DNS Zone Linking.
Breaking Changes
N/A
Testing Evidence
Evidence to be posted.
As part of this Pull Request I have
[X] Read the Contribution Guide and ensured this PR is compliant with the guide
[X] Ensured the resource API versions in .bicep file/s I am adding/editing are using the latest API version possible
[X] Checked for duplicate Pull Requests
[X] Associated it with relevant GitHub Issues
[X] (ALZ Bicep Core Team Only) Associated it with relevant ADO Items
[X] Ensured my code/branch is up-to-date with the latest changes in the main branch
[X] Performed testing and provided evidence.
[X] Updated one or more of the following tests (if required)
Unit
Linting
E2E (End-To-End)
ValidateAzCloud (Base validation in Azure Cloud)
ValidateMcCloud (Base validation in Azure China Cloud)
[X] Updated relevant and associated documentation (e.g. Contribution Guide, Module READMEs, Wiki Docs etc.)
[X] If relevant, created or updated Code Tours here
New module working - orchestration module test to be done
/azp run validateazcloud
/azp run e2e
/azp run validateazcloud
| gharchive/pull-request | 2023-03-13T15:55:37 | 2025-04-01T06:36:44.548504 | {
"authors": [
"JamJarchitect"
],
"repo": "Azure/ALZ-Bicep",
"url": "https://github.com/Azure/ALZ-Bicep/pull/476",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1221331356 | Add NRMS/Hypernet tags to cluster resource group in CI mode
Which issue this PR addresses:
The RP fails cluster installs if NSGs are attached to subnets. Simply Secure v1 applies NSGs in our subscriptions when these tags aren't present, resulting in a race condition.
What this PR does / why we need it:
Adds the tags needed to stop SSv1 from applying NSGs to subnets: https://eng.ms/docs/security-compliance-identity-and-management-scim/security/azure-security/security-health-analytics/network-isolation/netiso-teamdocs/netiso-program-overview/manage/hypernet-and-nrms-simply-secure-v1-network-security-rules
Test plan for issue:
Deploy this commit to central us euap (where the issue is faced)
Is there any documentation that needs to be updated for this PR?
No
/approve
hold for now as tagging the resource group didn't have the desired effect. We'll need to update these tags with a temporary SSv1 exception number and re-test.
@ross-bryan we can open a new PR with exception tags later if needed. Closing this one.
| gharchive/pull-request | 2022-04-29T17:34:44 | 2025-04-01T06:36:44.552595 | {
"authors": [
"cadenmarchese",
"rogbas"
],
"repo": "Azure/ARO-RP",
"url": "https://github.com/Azure/ARO-RP/pull/2096",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1615840762 | Fix Purge MSFT CI Subscription
Which issue this PR addresses:
Fixes
What this PR does / why we need it:
Test plan for issue:
Is there any documentation that needs to be updated for this PR?
Also fixes Purge Red Hat Dev-Gratis: https://msazure.visualstudio.com/AzureRedHatOpenShift/_build/results?buildId=69797295&view=logs&j=a4f1910f-c367-5697-edcd-724d81cde23b&t=2152b647-0197-5277-e95c-5d58fca9c7d3
| gharchive/pull-request | 2023-03-08T20:00:04 | 2025-04-01T06:36:44.555306 | {
"authors": [
"cadenmarchese"
],
"repo": "Azure/ARO-RP",
"url": "https://github.com/Azure/ARO-RP/pull/2766",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1922661909 | Set maintenace task to "" after PUCM pending operation
Which issue this PR addresses:
Fixes bug introduced by https://github.com/Azure/ARO-RP/pull/3194
What this PR does / why we need it:
We never unset PUCM pending maintenance task
If we do admin update with no maintenance task set, we're suppose to do a complete admin update, but instead we get PUCM pending task completed again
Test plan for issue:
Unit
INT
Is there any documentation that needs to be updated for this PR?
No - bug fix
/azp run e2e
| gharchive/pull-request | 2023-10-02T20:37:16 | 2025-04-01T06:36:44.558564 | {
"authors": [
"bennerv",
"niontive"
],
"repo": "Azure/ARO-RP",
"url": "https://github.com/Azure/ARO-RP/pull/3197",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1654669573 | refactor: lint errors; strings, undefinied nums, global vars
What type of PR is this?
What this PR does / why we need it:
Which issue(s) this PR fixes:
Fixes #
Requirements:
[ ] uses conventional commit messages
[ ] includes documentation
[ ] adds unit tests
[ ] tested upgrade from previous version
Special notes for your reviewer:
Release note:
none
Pull Request Test Coverage Report for Build 4612982352
44 of 126 (34.92%) changed or added relevant lines in 8 files are covered.
2 unchanged lines in 2 files lost coverage.
Overall coverage decreased (-2.3%) to 77.079%
Changes Missing Coverage
Covered Lines
Changed/Added Lines
%
pkg/agent/utils.go
9
13
69.23%
pkg/agent/datamodel/types.go
12
17
70.59%
pkg/agent/datamodel/sig_config.go
0
73
0.0%
Files with Coverage Reduction
New Missed Lines
%
pkg/agent/datamodel/sig_config.go
1
42.29%
pkg/agent/datamodel/types.go
1
80.43%
Totals
Change from base Build 4612232559:
-2.3%
Covered Lines:
2011
Relevant Lines:
2609
💛 - Coveralls
| gharchive/pull-request | 2023-04-04T22:40:39 | 2025-04-01T06:36:44.571438 | {
"authors": [
"AlisonB319",
"coveralls"
],
"repo": "Azure/AgentBaker",
"url": "https://github.com/Azure/AgentBaker/pull/2976",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2191269333 | docs: add Windows image release notes for 2024-3b
What type of PR is this?
/kind documentation
What this PR does / why we need it:
add Windows image release notes for 2024-3b. Reference: #4175
Which issue(s) this PR fixes:
Fixes #
Requirements:
[x] uses conventional commit messages
[ ] includes documentation
[ ] adds unit tests
[ ] tested upgrade from previous version
Special notes for your reviewer:
Release note:
none
Pull Request Test Coverage Report for Build 8321780001
Details
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 79.343%
Totals
Change from base Build 8304229022:
0.0%
Covered Lines:
2293
Relevant Lines:
2890
💛 - Coveralls
| gharchive/pull-request | 2024-03-18T05:31:51 | 2025-04-01T06:36:44.579595 | {
"authors": [
"coveralls",
"junjiezhang1997"
],
"repo": "Azure/AgentBaker",
"url": "https://github.com/Azure/AgentBaker/pull/4183",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
496001586 | Event grid subscriptions no longer working
I had created some EventGrid subscriptions to a preview app configuration instance in East US. Today I can no longer see the subscription and receive the following error when clicking on Events
Are event grid subscriptions really no longer available in East US or is this a UI bug?
@nhart12 This is a UI bug, working on resolving it. Thanks for notifying us!
This has been resolved.
Thanks @jimmyca15 the UI seems to work now, the events however are not firing consistently in our app configuration instance
@nhart12. Is this happening in a single region with a single store? Do you mean events are firing sometimes, and then not firing other times or that they're not firing at all?
@jimmyca15 So this is happening in our East US store. Just tried again this morning, created a couple settings, updated them several times and then deleted. I have an EventGrid subscription setup for key-value Modified and Deleted. Only the deleted events fired
There was a bug in event grid publishing identified last week. We are in the process of rolling out our fix for that. That may be causing your issue. When the fix is rolled out I will let you know.
Ok perfect thanks for the info @jimmyca15 !
@jimmyca15 Can you provide an update regarding when you anticipate the event grid publishing issues may be resolved? Also, is a GA estimate available...the most recent discussion around this mentioned October timeframe, is that still a possibility?
@jmsalvo event grid publishing changes are getting rolled out region by region. Expected to all regions by thursday eod.
Estimated GA is still end of October. Thanks.
Yesterday we deployed a fix to event grid to improve some event publishing reliability issues we were facing. Please open another issue if you are still experiencing problems.
Hi @nhart12 and @jmsalvo, I want to give you an update. We just posted our GA and pricing plan. Please check it out.
https://github.com/Azure/AppConfiguration-Announcements/issues/2
| gharchive/issue | 2019-09-19T20:07:16 | 2025-04-01T06:36:44.584928 | {
"authors": [
"jimmyca15",
"jmsalvo",
"nhart12",
"zhenlan"
],
"repo": "Azure/AppConfiguration",
"url": "https://github.com/Azure/AppConfiguration/issues/142",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1759342230 | 🪲 Bug Report - Upgrade Fails for SF VMSS with KeyVault Extension
Describe the bug
Upgrade fails due to KV VM extension timeout:
2023-06-14T21:07:57+00 [Information]:############################## Initializing Start-AzBasicLoadBalancerUpgrade ##############################
2023-06-14T21:07:57+00 [Information]:[Start-AzBasicLoadBalancerUpgrade] PowerShell Version: **7.3.4**
2023-06-14T21:07:57+00 [Information]:[Start-AzBasicLoadBalancerUpgrade] AzureBasicLoadBalancerUpgrade **Version: 2.0.19**
...
2023-06-14T21:08:00+00 [Information]:[Test-SupportedMigrationScenario] Checking whether VMSS scale set 'quotavmssdevbn' is a Service Fabric cluster...
WARNING: 2023-06-14T21:08:00+00 [Warning]:[Test-SupportedMigrationScenario] **VMSS appears to be a Service Fabric** cluster based on extension profile. SF Clusters experienced potentically significant downtime during migration using this PowerShell module. In testing, a 5-node Bronze cluster was unavailable for about 30 minutes and a 5-node Silver cluster was unavailabile for about 45 minutes. Shutting down the cluster VMSS prior to initiating migration will result in a more consistent experience of about 5 minutes to complete the LB migration. For Service Fabric clusters that require minimal / no connectivity downtime, adding a new nodetype with standard load balancer and IP resources is a better solution.
Do you want to proceed with the migration of your Service Fabric Cluster's Load Balancer?
...
2023-06-14T21:08:00+00 [Information]:[PublicLBMigration] **Public Load Balancer Detected**. Initiating Public Load Balancer Migration
...
2023-06-14T21:24:34+00 [Information]:[NatRulesMigration] Waiting for saving standard load balancer LB-quota-cluster-dev-bn job to complete...
2023-06-14T21:24:34+00 [Information]:[NatRulesMigration] Nat Rules Migration Completed
2023-06-14T21:24:34+00 [Information]:[InboundNatPoolsMigration] Initiating Inbound NAT Pools Migration
2023-06-14T21:24:34+00 [Information]:[InboundNatPoolsMigration] Adding Inbound NAT Pool LoadBalancerBEAddressNatPool to Standard Load Balancer
2023-06-14T21:24:34+00 [Information]:[InboundNatPoolsMigration] Saving Standard Load Balancer LB-quota-cluster-dev-bn
2023-06-14T21:24:49+00 [Information]:[InboundNatPoolsMigration] Waiting for saving standard load balancer LB-quota-cluster-dev-bn job to complete...
2023-06-14T21:24:49+00 [Information]:[GetVmssFromBasicLoadBalancer] Initiating GetVmssFromBasicLoadBalancer
2023-06-14T21:24:49+00 [Information]:[GetVmssFromBasicLoadBalancer] Getting VMSS object '/subscriptions/.../resourcegroups/azure-quota-dev-eastus2/providers/microsoft.compute/virtualmachinescalesets/quotavmssdevbn' from Azure
2023-06-14T21:24:49+00 [Information]:[GetVmssFromBasicLoadBalancer] VMSS loaded Name quotavmssdevbn from RG azure-quota-dev-eastus2
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Adding InboundNATPool to VMSS quotavmssdevbn
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Checking if VMSS 'quotavmssdevbn' NIC 'NIC-azure-quota-dev-eastus2' IPConfig 'NIC-azure-quota-dev-eastus2' should be associated with NAT Pool 'LoadBalancerBEAddressNatPool'
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Adding NAT Pool 'LoadBalancerBEAddressNatPool' to IPConfig 'NIC-azure-quota-dev-eastus2'
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Migrate NetworkInterface Configurations completed
2023-06-14T21:24:49+00 [Information]:[InboundNatPoolsMigration] Saving VMSS quotavmssdevbn
2023-06-14T21:24:49+00 [Information]:[UpdateVmss] Updating configuration of VMSS 'quotavmssdevbn'
2023-06-14T21:25:04+00 [Information]:[UpdateVmss] Waiting for job (id: '5') updating VMSS 'quotavmssdevbn' to complete...
...
2023-06-14T23:10:50+00 [Information]:[UpdateVmss] Waiting for job (id: '5') updating VMSS 'quotavmssdevbn' to complete...
InvalidOperation: Long running operation failed with status 'Failed'. Additional Info:'Provisioning of VM extension **KvVmExtension** has timed out. Extension provisioning has taken too long to complete. The extension did not report a message. More information on troubleshooting is available at https://aka.ms/vmextensionwindowstroubleshoot'
ErrorCode: VMExtensionProvisioningTimeout
ErrorMessage: Provisioning of VM extension KvVmExtension has timed out. Extension provisioning has taken too long to complete. The extension did not report a message. More information on troubleshooting is available at https://aka.ms/vmextensionwindowstroubleshoot
ErrorTarget: 0
StartTime: 6/14/2023 9:24:52 PM
EndTime: 6/14/2023 11:10:27 PM
OperationID: 85ee53b5-9ce3-4458-9edd-f46e8c7baf02
Status: Failed
Write-Error: 2023-06-14T23:10:50+00 [Error]:[InboundNatPoolsMigration] An error occured when attempting to update VMSS network config on the new Standard LB backend pool membership. To recover address
the following error, and try again specifying the -FailedMigrationRetryFilePath parameter and Basic Load Balancer backup State file located either in this directory or the directory
specified with -RecoveryBackupPath
To Reproduce
Steps to reproduce the behavior:
VMSS
Public LB
KVVMExtension [this case, extension adds a cert to local store, auto upgrade disabled]
SF Cluster [?]
Additional context - please include:
See log
Trying to test this scenario, I can't get the LB migration to even start due to:
InvalidOperation: Long running operation failed with status 'Failed'. Additional Info:'VM has reported a failure when processing extension 'vmss-01'. Error message: "Failed to parse the configuration settings with: 'not a string'"
More information on troubleshooting is available at https://aka.ms/vmextensionwindowstroubleshoot '
ErrorCode: VMExtensionProvisioningError
ErrorMessage: VM has reported a failure when processing extension 'vmss-01'. Error message: "Failed to parse the configuration settings with: 'not a string'"
This issue is reportedly intermittent...still working to repro
Closing due to lack of activity and reproducibility
@mbrat2005 Im experiencing the same issue. Did you ever find a solution?
Hi Andrew,
I haven't made progress on this one, since I couldn't seem to repro it. Would you be able to share your upgrade log for details? Also, are you upgrading a basic LB for a Service Fabric Cluster?
Thanks!
Matthew
From: Andrew Smith @.>
Sent: Wednesday, July 3, 2024 13:29
To: Azure/AzLoadBalancerMigration @.>
Cc: Mention @.>; Author @.>; Comment @.>; Assign @.>; State change @.***>
Subject: Re: [Azure/AzLoadBalancerMigration] 🪲 Bug Report - Upgrade Fails for SF VMSS with KeyVault Extension (Issue #65)
@mbrat2005https://github.com/mbrat2005 Im experiencing the same issue. Did you ever find a solution?
—
Reply to this email directly, view it on GitHubhttps://github.com/Azure/AzLoadBalancerMigration/issues/65#issuecomment-2207038571 or unsubscribehttps://github.com/notifications/unsubscribe-auth/AGBW6WHTSMOHXES4YPWH4F3ZKRGLFBFKMF2HI4TJMJ2XIZLTSWBKK5TBNR2WLJDUOJ2WLJDOMFWWLO3UNBZGKYLEL5YGC4TUNFRWS4DBNZ2F6YLDORUXM2LUPGBKK5TBNR2WLJDUOJ2WLJDOMFWWLLTXMF2GG2C7MFRXI2LWNF2HTAVFOZQWY5LFUVUXG43VMWSG4YLNMWVXI2DSMVQWIX3UPFYGLAVFOZQWY5LFVI2DKNRTHEYTONZQHGSG4YLNMWUWQYLTL5WGCYTFNSBKK5TBNR2WLKRVGU2DSOJZGYZDQM5ENZQW2ZNJNBQXGX3MMFRGK3FMON2WE2TFMN2F65DZOBS2YSLTON2WKQ3PNVWWK3TUUZ2G64DJMNZZJAVEOR4XAZNKOJSXA33TNF2G64TZUV3GC3DVMWUTKMZZHE4TSOJYG6BKI5DZOBS2K2LTON2WLJLWMFWHKZNKGE3TKOJTGQZDEMZQQKSHI6LQMWSWYYLCMVWKK5TBNR2WLKRUGU3DGOJRG43TAOMCUR2HS4DFUVWGCYTFNSSXMYLMOVS2UNJVGQ4TSOJWGI4DHJ3UOJUWOZ3FOKTGG4TFMF2GK.
You are receiving this email because you were mentioned.
Triage notifications on the go with GitHub Mobile for iOShttps://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675 or Androidhttps://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub.
@AndrewCS149 I haven't made progress on this one, since I couldn't seem to repro it. Would you be able to share your upgrade log for details? Also, are you upgrading a basic LB for a Service Fabric Cluster?
| gharchive/issue | 2023-06-15T18:26:48 | 2025-04-01T06:36:44.596575 | {
"authors": [
"AndrewCS149",
"mbrat2005"
],
"repo": "Azure/AzLoadBalancerMigration",
"url": "https://github.com/Azure/AzLoadBalancerMigration/issues/65",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2210379760 | Error while clicking Manage button from the already installed BitSight solution under the Content Hub
Installed the BitSight solution from the Content Hub from the azure portal. After successfully installed when I go to manage it by clicking on the Manage button from the installed solution page from the Content Hub getting the "not found" error as below:
To Reproduce
Steps to reproduce the behavior:
Login to Azure Portal
Go to Microsoft Sentinel Workspace
Go to the Content Hub from the side menu and search for the BitSight solution
Install that solution by providing the required information of Subscription, Resource Group etc...
Again go to the BitSight solution from the content hub and click on the Manage button
Expected behavior
When clicking on the Manage button from the BitSight solution under the content hub it should display BitSight Solution Manage screen which all the available component of that solution.
Desktop (please complete the following information):
OS: Windows
Browser: chrome
Version: 122.0.6261.129 (Official Build) (64-bit)
Hi @jayeshprajapaticrest,
Thanks for flagging this issue, we will investigate this issue and get back to you with some updates by 03 Apr 2024. Thanks!
@v-rusraut
Understood. Thanks
Hi @jayeshprajapaticrest,
Please check with below link , if you still facing issue,please let us know.
https://ms.portal.azure.com/
Thanks
@v-rusraut
Thanks
Still getting the same error.
@v-rusraut
I checked today and the issue get resolved.
Thanks
Just to know Can you please let me know the cause of this issue?
| gharchive/issue | 2024-03-27T10:10:15 | 2025-04-01T06:36:44.609120 | {
"authors": [
"jayeshprajapaticrest",
"v-rusraut"
],
"repo": "Azure/Azure-Sentinel",
"url": "https://github.com/Azure/Azure-Sentinel/issues/10208",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1435515130 | Create Theom Solution
Required items, please complete
Change(s):
Adding Solution for Theom
Added 1 Workbook, 20 Analytics Rules
Reason for Change(s):
New
Version Updated:
Yes
Testing Completed:
Yes
Checked that the validations are passing and have addressed any issues that are present:
Yes
@theomsoc :
Logo is looking good
Workbook is looking good
Could you please create a solution package by using V2 tool
Please follow the below URL and do the package creation. Thanks!
Azure-Sentinel/README.md at master · Azure/Azure-Sentinel (github.com)
@v-spadarthi should I use my branch as input for Logo and Basepath in input file at https://github.com/Azure/Azure-Sentinel/blob/master/Tools/Create-Azure-Sentinel-Solution/V2/README.md#create-input-file
@v-spadarthi added solution package, please review
@v-spadarthi thanks for the review. Addressed comments and rebuilt package. Shortcut names and atypical are as expected, to match the wording on our application.
@theomsoc : Data connector is looking fine
CreateUI definition
Main Template
Arm-ttk also running fine
@aprakash13 : Please review the analytics and provide your signoff. Thanks!
Once Analytics rules reviewed by @aprakash13 we are going to approve this PR.
@aprakash13 addressed comments and re-published new package
@aprakash13 @v-spadarthi could you please review and merge PR soon, as we are waiting on publishing solution.
Thanks @aprakash13 !! @v-spadarthi please do the needful.
@theomsoc : Please add the solution input file for that create Data folder under Solutions/Theom/ and keep it here.
@theomsoc : Thanks for adding solution input file.
Plz remove the empty properties from solution input file like parsers and playbooks
@theomsoc : Please add the solution input file for that create Data folder under Solutions/Theom/ and keep it here.
@v-spadarthi Added input file, resolved conflicts and removed empty properties.
If all looks good, please merge PR, thanks!
@v-spadarthi could you please check and merge PR soon, so we are able to publish today ?
Create UI definition is looking good
Arm ttk also running fine
All are looking good
Approved and merged since Suresh's has approved the PR
| gharchive/pull-request | 2022-11-04T04:01:59 | 2025-04-01T06:36:44.621382 | {
"authors": [
"theomsoc",
"v-dvedak",
"v-spadarthi"
],
"repo": "Azure/Azure-Sentinel",
"url": "https://github.com/Azure/Azure-Sentinel/pull/6556",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
216154443 | Add-GalleryItems from local folder; Get-GalleryItem; Remove-GalleryIt…
Add-GalleryItems from local folder; Get-GalleryItem; Remove-GalleryItem in 1 script
@larisaborodina,
Thanks for your contribution as a Microsoft full-time employee or intern. You do not need to sign a CLA.
Thanks,
Microsoft Pull Request Bot
| gharchive/pull-request | 2017-03-22T17:49:38 | 2025-04-01T06:36:44.627214 | {
"authors": [
"larisaborodina",
"msftclas"
],
"repo": "Azure/AzureStack-Tools",
"url": "https://github.com/Azure/AzureStack-Tools/pull/122",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2435588234 | Update script ERROR
Describe the bug
Run the update script to update existing SaaS-Accelerator in Azure, but get error
To Reproduce
Steps to reproduce the behavior:
Fork the SaaS-Accelerator code with latest release 7.6.3
Logon to Azure Portal > Open Cloud Shell and run below update-script
wget https://dotnet.microsoft.com/download/dotnet/scripts/v1/dotnet-install.sh; chmod +x dotnet-install.sh;
./dotnet-install.sh -version 6.0.423; $ENV:PATH="$HOME/.dotnet:$ENV:PATH";
dotnet tool install --global dotnet-ef --version 6.0.1; git clone https://github.com/xxxx/Commercial-Marketplace-SaaS-Accelerator.git; cd ./Commercial-Marketplace-SaaS-Accelerator/deployment;
.\Upgrade.ps1 -WebAppNamePrefix "xxxxx"
-ResourceGroupForDeployment "xxxx"
See the errors:
Line |
32 | Param(
| ~~~~~
| The term 'Param' is not recognized as a name of a cmdlet, function, script file, or executable program. Check the spelling of the name, or if a path was
| included, verify that the path is correct and try again.
Deploying new database
ERROR: argument --vault-name: expected one argument
String-Between: /home/martel/Commercial-Marketplace-SaaS-Accelerator/deployment/Upgrade.ps1:64
Line |
64 | $Server = String-Between -source $ConnectionString -start "Data Sourc …
| ~~~~~~~~~~~~~~~~~
| Cannot bind argument to parameter 'Source' because it is an empty string.
Expected behavior
Expect to see successful message
Environment (please complete the following information):
Are you using the CloudShell? YES
If not using the CloudShell:
OS: [e.g. iOS] Window
PowerShell version
Modules installed: Pre installed SaaS-Accelerator 6.5
PowerShell Version Used: 7.4.3
The latest PR should have fixed this. @code4clouds can you see if it’s still an issue please
I got the fix and rerun the Upgrade.ps1 script, but get error with MSI:
Deploying new database
ERROR: Failed to connect to MSI. Please make sure MSI is configured correctly.
Get Token request returned: <Response [400]>
Upgrade.ps1:
Line |
8 | .\Upgrade.ps1 `
| ~~~~~~~~~~~~~~~
| Cannot bind argument to parameter 'Source' because it is an empty string.
@marteltran if you have already installed latest any reason why you are upgrading?
But if you are anyway, please make sure to remove vnet protection for sql, keyvault and app services during the upgrade and add it back again after the upgrade
I have installed the old version (6.5.x) so I want to upgrade to the latest (7.4.3).
Just confirm:
Do I have to DELETE the key-vault and 2 App services (Admin & Portal) before running the upgrade script? (I do not use VNET).
How to add the key-vault after all?
@marteltran - lets take a step back, but the latest is 7.6.2 not 7.4.3. I dont even see 7.4.3 anywhere am I missing something?
Also there is no comment about deleting anything, so please dont delete anything unless you are aware of what you are doing.
I confuse. When I fork the main branch, What is current version? and How to make sure that my fork contain 7.6.2?
| gharchive/issue | 2024-07-29T14:30:04 | 2025-04-01T06:36:44.641187 | {
"authors": [
"marteltran",
"santhoshb-msft"
],
"repo": "Azure/Commercial-Marketplace-SaaS-Accelerator",
"url": "https://github.com/Azure/Commercial-Marketplace-SaaS-Accelerator/issues/740",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
476450890 | 'BadEncodingLimitsExceeded ' received when creating subscriptions
We have OPC Publisher 2.3.3 running on iotedge 1.0.6.1 (3fa6cbef8b7fc3c55a49a622735eb1021b8a5963)
On startup up of the Publisher we are getting a 'BadEncodingLimitsExceeded ' error.
How should we go about debugging/resolving this issue?
[04:34:04 INF] Publisher is running. Press CTRL-C to quit.
[04:34:04 INF] Connect and monitor session and nodes on endpoint 'opc.tcp://10.70.64.95:49320'.
[04:34:06 INF] Create unsecured session for endpoint URI 'opc.tcp://10.70.64.95:49320' with timeout of 10000 ms.
[04:34:06 INF] Session successfully created with Id ns=1;g=73c5b05f-456f-49ec-a62b-a555657aaa6d.
[04:34:06 INF] The session to endpoint 'opc.tcp://10.70.64.95:49320/' has 3 entries in its namespace array:
[04:34:06 INF] Namespace index 0: http://opcfoundation.org/UA/
[04:34:06 INF] Namespace index 1: urn:NZS070HSTOPC01X.prod01.inroot.net:PTC.ThingWorxIC.1:UA%20Server
[04:34:06 INF] Namespace index 2: ThingWorx Industrial Connectivity
[04:34:06 INF] Created subscription with id 8 on endpoint 'opc.tcp://10.70.64.95:49320'
[04:34:06 INF] Create subscription on endpoint 'opc.tcp://10.70.64.95:49320' requested OPC publishing interval is 30000 ms. (revised: 30000 ms)
[04:34:06 INF] Start monitoring items on endpoint 'opc.tcp://10.70.64.95:49320'. Currently monitoring 0 items.
[04:34:06 INF] Now monitoring 10000 items in subscription with id '8'
[04:34:07 INF] Now monitoring 20000 items in subscription with id '8'
[04:34:08 INF] Now monitoring 30000 items in subscription with id '8'
[04:34:10 INF] Now monitoring 40000 items in subscription with id '8'
[04:34:14 INF] Now monitoring 50000 items in subscription with id '8'
[04:34:14 ERR] Unhandled OPC UA ServiceResultException 'BadEncodingLimitsExceeded 'Could not send request to server.'' when monitoring node 'ns=2;s=NZ070_Darfield_NZS070SRV01X.C270.NZ070_B05VA0506_AlmFault' on endpoint 'opc.tcp://10.70.64.95:49320'. Continue.
[04:34:14 INF] Now monitoring 60000 items in subscription with id '8'
[04:34:14 ERR] Unhandled OPC UA ServiceResultException 'BadEncodingLimitsExceeded 'Could not send request to server.'' when monitoring node 'ns=2;s=NZ070_Darfield_NZS070SRV01X.C801.NZ070_S1PW1P01_Msg1' on endpoint 'opc.tcp://10.70.64.95:49320'. Continue.
[04:34:15 ERR] Exception
Opc.Ua.ServiceResultException: Could not send request to server.
at Opc.Ua.Bindings.ChannelAsyncOperation`1.End(Int32 timeout, Boolean throwOnError)
at Opc.Ua.Bindings.UaSCUaBinaryClientChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.SendRequest(IServiceRequest request)
at Opc.Ua.SessionClient.CreateMonitoredItems(RequestHeader requestHeader, UInt32 subscriptionId, TimestampsToReturn timestampsToReturn, MonitoredItemCreateRequestCollection itemsToCreate, MonitoredItemCreateResultCollection& results, DiagnosticInfoCollection& diagnosticInfos)
at Opc.Ua.Client.Subscription.CreateItems()
at Opc.Ua.Client.Subscription.ApplyChanges()
at OpcPublisher.OpcSession.MonitorNodesAsync(CancellationToken ct) in /app/opcpublisher/OpcSession.cs:line 561
[04:34:25 INF] Start monitoring items on endpoint 'opc.tcp://10.70.64.95:49320'. Currently monitoring 66541 items.
[04:34:25 ERR] Unhandled OPC UA ServiceResultException 'BadEncodingLimitsExceeded 'Could not send request to server.'' when monitoring node 'ns=2;s=NZ070_Darfield_NZS070SRV01X.C270.NZ070_B05VA0506_AlmFault' on endpoint 'opc.tcp://10.70.64.95:49320'. Continue.
[04:34:25 ERR] Exception
Opc.Ua.ServiceResultException: Could not send request to server.
at Opc.Ua.Bindings.ChannelAsyncOperation`1.End(Int32 timeout, Boolean throwOnError)
at Opc.Ua.Bindings.UaSCUaBinaryClientChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.SendRequest(IServiceRequest request)
at Opc.Ua.SessionClient.CreateMonitoredItems(RequestHeader requestHeader, UInt32 subscriptionId, TimestampsToReturn timestampsToReturn, MonitoredItemCreateRequestCollection itemsToCreate, MonitoredItemCreateResultCollection& results, DiagnosticInfoCollection& diagnosticInfos)
at Opc.Ua.Client.Subscription.CreateItems()
at Opc.Ua.Client.Subscription.ApplyChanges()
at OpcPublisher.OpcSession.MonitorNodesAsync(CancellationToken ct) in /app/opcpublisher/OpcSession.cs:line 561
I have tried removing the tags that are reporting errors but it appears to just report the next tag as error'd after this.
This error is also mentioned on 2 other OPC github projects.
https://github.com/OPCFoundation/UA-.NET-Legacy/issues/107
https://github.com/open62541/open62541/issues/434
Neither is very clear to me why this might be happening, but sounds like it could be something going over the size limit for an OPCUA message?
@marcschier Yes, this use case is working now that we have the correct path.
We are seeing some performance challenges, I think it would be good to consider batching the OPC UA subscriptions if the resulting OPC UA group would result in a really large subscription. Would be interesting to see if this would improve performance or reliability.
| gharchive/issue | 2019-04-30T05:02:14 | 2025-04-01T06:36:44.649221 | {
"authors": [
"jmbrunskill"
],
"repo": "Azure/Industrial-IoT",
"url": "https://github.com/Azure/Industrial-IoT/issues/55",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2493107915 | WYSIWYG Formatting Issues
Describe the Bug with repro steps
Update any WYSIWYG text fields in new Logic App Editor
Save
On reload, line breaks of edited fields tend to have been removed
In my organization, this has been observed especially with Outlook Email "Message" field and Azure DevOps > Create a Work Item "Description" field
What type of Logic App Is this happening in?
Consumption (Portal)
Which operating system are you using?
Windows, Linux
Are you using new designer or old designer
New Designer
Did you refer to the TSG before filing this issue? https://aka.ms/lauxtsg
Yes
Workflow JSON
No response
Screenshots or Videos
Formatting before saving changes
Formatting after saving changes
Browser
Microsoft Edge, Chrome, Firefox
Additional context
This has occurred in 4 of our logic apps that I'm aware of
This has occurred under at least two different users on two different computers
This seems to generally occur on larger logic apps
Only occurs on New Designer, remediated by using Old Designer, which is slated for retirement
Also confirmed issue on Firefox from Fedora Linux
This has been fixed but not deployed, will start rolling out deployment for this on Monday https://github.com/Azure/LogicAppsUX/pull/5392
Right on, thanks a ton for the heads up, I'll look forward to it!
Actually Tuesday! Monday holiday in US
In case anyone finds this, deployment to prod took effect today 9/10/2024
Thanks a ton for all of your help!
| gharchive/issue | 2024-08-28T22:48:18 | 2025-04-01T06:36:44.658061 | {
"authors": [
"DanielleCogs",
"DrLeeroyPhD"
],
"repo": "Azure/LogicAppsUX",
"url": "https://github.com/Azure/LogicAppsUX/issues/5544",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
233733124 | Update Kubernetes Spark Install Demo
Update the Kubernetes + Spark install demo to work with the SimDem update. Adds cloning the Kubernetes GitHub repository to the home directory as part of the preparation as it takes a while.
@suhaildawood,
Thanks for your contribution as a Microsoft full-time employee or intern. You do not need to sign a CLA.
Thanks,
Microsoft Pull Request Bot
| gharchive/pull-request | 2017-06-05T22:52:18 | 2025-04-01T06:36:44.659909 | {
"authors": [
"msftclas",
"suhaildawood"
],
"repo": "Azure/acs-demos",
"url": "https://github.com/Azure/acs-demos/pull/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
225463512 | Outbound internet does not work with CNI plugin
To repro:
deploy a cluster with CNI
kubectl exec to the container, and try to access the internet from a container
@ofiliz can you take a look?
I propose that we revise the issue title: "Kubernetes pods do not have outbound internet access with CNI plugin in l2tunnel mode".
Azure-vnet CNI/CNM plugins in regular IaaS VMs can be given outgoing access to internet with a simple NAT rule:
iptables -t nat -A POSTROUTING -m iprange ! --dst-range 168.63.129.16 -m addrtype ! --dst-type local ! -d {{{vnetCidr}}} -j MASQUERADE
The problem is the traffic policy enforced by CNI plugin and iptables rules set by kube-proxy are incompatible. We'll have to add this capability directly to Azure SDN stack, instead of SNAT'ing in the VM.
Discussed further with @seanknox and it is better to remove the examples until this can be fixed, otherwise customers will keep hitting this error.
Verified this is no longer an issue. On a cluster with azure CNI enabled:
$ ps auxfww | grep '/etc/cni/'
azureus+ 19583 0.0 0.0 12948 936 pts/0 S+ 00:06 0:00 \_ grep --color=auto /etc/cni/
root 6778 0.0 0.2 161524 17400 ? Ssl Nov21 0:00 /usr/bin/docker run --net=host --pid=host --privileged --rm --volume=/dev:/dev --volume=/sys:/sys:ro --volume=/var/run:/var/run:rw --volume=/var/lib/docker/:/var/lib/docker:rw --volume=/var/lib/kubelet/:/var/lib/kubelet:shared --volume=/var/log:/var/log:rw --volume=/etc/kubernetes/:/etc/kubernetes:ro --volume=/srv/kubernetes/:/srv/kubernetes:ro --volume=/etc/cni/:/etc/cni:ro --volume=/opt/cni/:/opt/cni:ro --volume=/var/lib/waagent/ManagedIdentity-Settings:/var/lib/waagent/ManagedIdentity-Settings:ro gcrio.azureedge.net/google_containers/hyperkube-amd64:v1.7.9 /hyperkube kubelet --kubeconfig=/var/lib/kubelet/kubeconfig --require-kubeconfig --pod-infra-container-image=gcrio.azureedge.net/google_containers/pause-amd64:3.0 --address=0.0.0.0 --allow-privileged=true --enable-server --pod-manifest-path=/etc/kubernetes/manifests --cluster-dns=10.0.0.10 --cluster-domain=cluster.local --node-labels=kubernetes.io/role=master,kubernetes.azure.com/cluster=francis-1dot6 --cloud-provider=azure --cloud-config=/etc/kubernetes/azure.json --azure-container-registry-config=/etc/kubernetes/azure.json --network-plugin=cni --max-pods=30 --node-status-update-frequency=10s --image-gc-high-threshold=85 --image-gc-low-threshold=80 --v=2 --non-masquerade-cidr=10.0.0.0/8 --register-node=true --register-with-taints=node-role.kubernetes.io/master=true:NoSchedule
Was able to create a container, apt-get install curl, and curl to the internet:
$ kubectl exec -it shell-demo -- /bin/bash
root@shell-demo:/# curl google.com
<HTML><HEAD><meta http-equiv="content-type" content="text/html;charset=utf-8">
<TITLE>301 Moved</TITLE></HEAD><BODY>
<H1>301 Moved</H1>
The document has moved
<A HREF="http://www.google.com/">here</A>.
</BODY></HTML>
Awesome! Do you know what changed? Newer CNI image?
@seanknox I suspect it was the addition of a masquerade rule in iptables. See here for an example:
https://github.com/Azure/acs-engine/blob/master/parts/kubernetesmastercustomdata.yml#L283
@jackfrancis @seanknox Yes, it was fixed with the IPTables rule. Internet access should be working.
| gharchive/issue | 2017-05-01T16:12:15 | 2025-04-01T06:36:44.665397 | {
"authors": [
"anhowe",
"jackfrancis",
"ofiliz",
"seanknox",
"sharmasushant"
],
"repo": "Azure/acs-engine",
"url": "https://github.com/Azure/acs-engine/issues/561",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2398197922 | Drop Azure Arc ESU Dashboard
Add in a drop referencing an Azure Arc Windows ESU Dashboard that people can use to view their Arc and ESU status.
Thanks for the feedback @fcabrera23, I've made the necessary changes. Let me know your thoughts. Thanks!
| gharchive/pull-request | 2024-07-09T13:15:45 | 2025-04-01T06:36:44.666600 | {
"authors": [
"weeyin83"
],
"repo": "Azure/arc_jumpstart_drops",
"url": "https://github.com/Azure/arc_jumpstart_drops/pull/119",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
981078307 | Redeploying Bastion via template fails (Subnet AzureBastionSubnet is in use)
I'm not sure if this belongs here, but I'll give it a try.
I've created a bicep template to deploy a vnet resource with a Bastion resource. The first time deploying this template works as expected and all resources are created.
But when redeploying the template, I get the error "Subnet AzureBastionSubnet is in use by /subscriptions/..../resourceGroups/rg-test/providers/Microsoft.Network/bastionHosts/avd-vnet-bastion/bastionHostIp
Configurations/IpConf and cannot be deleted. In order to delete the subnet, delete all the resources within the subnet."
What I'm expecting is that when redeploying the template, it would detected that the resources exist and no changes are made to the resources. This is how other templates work when deploying other resources. I don't understand why it's trying to delete the AzureBastionSubnet.
See attached file that is the bicep template.
bastion.txt
@nielsvdc This is related to this behavior stated here: https://stackoverflow.com/questions/55901747/azure-arm-code-inusesubnetcannotbedeleted-when-trying-to-update-vnet
When the vnet gets deployed and no subnets are specified in the properties section, the vnet tries to delete the subnets. However, services are connected and hence, the deployment fails. Just the first deployment works fine, as there is no subnet that is in use.
Sorry people, I was unavailable for a moment. I tested with redeployment with CLI and also via Azure GUI. These also both failed with error message "Subnet AzureBastionSubnet is in use by /subscriptions/.../resourceGroups/test/providers/Microsoft.Network/bastionHosts/avd-vnet-bastion-eus/bastionHostIpConfigurations/IpConf and cannot be deleted. In order to delete the subnet, delete all the resources within the subnet."
Hi @nielsvdc, if we try to run az deployment group what-if ... we get the below outcome which clearly shows that they are trying to delete IpConf. So this looks like an expected behaviour.
This is a possibility when different resources doesnt have context about other resources as explained in this article
If you still feel this should not be an expected behaviour then as we use azure-cli internally yo make these calls, they will be able to help you out. Can you please raise an issue here? https://github.com/Azure/azure-cli
Note: The result may contain false positive predictions (noise).
You can help us improve the accuracy of the result by opening an issue here: https://aka.ms/WhatIfIssues.
Resource and property changes are indicated with these symbols:
- Delete
~ Modify
= Nochange
* Ignore
The deployment will update the following scope:
Scope: /subscriptions/c00d16c7-6c1f-4c03-9be1-6934a4c49682/resourceGroups/deepak-rg
~ Microsoft.Network/bastionHosts/avd-vnet-bastion [2021-02-01]
- properties.dnsName: "bst-fffc8494-9771-467e-979f-389868569805.bastion.azure.com"
- properties.scaleUnits: 2
~ properties.ipConfigurations: [
~ 0:
- id: "/subscriptions/c00d16c7-6c1f-4c03-9be1-6934a4c49682/resourceGroups/deepak-rg/providers/Microsoft.Network/bastionHosts/avd-vnet-bastion/bastionHostIpConfigurations/IpConf"
]
~ sku.name: "Basic" => "Standard"
~ Microsoft.Network/publicIPAddresses/avd-vnet-bastion-ip [2021-02-01]
- sku.tier: "Regional"
= Microsoft.Network/virtualNetworks/myVnet [2021-02-01]
= Microsoft.Network/virtualNetworks/myVnet/subnets/AzureBastionSubnet [2021-02-01]
* Microsoft.Network/networkSecurityGroups/NRMS-mqpdnhjdzis7mmyVnet
* Microsoft.Network/networkSecurityGroups/NRMS-wlg2nxyjuzp7ymyVnet-AzureBastionSubnet
* Microsoft.Storage/storageAccounts/deepak21
* Microsoft.Storage/storageAccounts/deepak2121
* Microsoft.Storage/storageAccounts/devstore4z32rb65cghuc
* Microsoft.Web/serverFarms/AppServicePlan-GEN-UNIQUE
* Microsoft.Web/serverFarms/devplan
* Microsoft.Web/sites/devapp4z32rb65cghuc
Resource changes: 2 to modify, 2 no change, 8 to ignore.
Hi @nielsvdc are you still facing this issue? If no, then should we close this?
| gharchive/issue | 2021-08-27T09:57:15 | 2025-04-01T06:36:44.672960 | {
"authors": [
"marvinbuss",
"nielsvdc",
"t-dedah"
],
"repo": "Azure/arm-deploy",
"url": "https://github.com/Azure/arm-deploy/issues/80",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1323839318 | Ensure base class is absrtact in polymorphism
Description
Regen PR in sdk repo: https://github.com/Azure/azure-sdk-for-net/pull/30235
Checklist
To ensure a quick review and merge, please ensure:
[ ] The PR has a understandable title and description explaining the why and what.
[ ] The PR is opened in draft if not ready for review yet.
If opened in draft, please allocate sufficient time (24 hours) after moving out of draft for review
[ ] The branch is recent enough to not have merge conflicts upon creation.
Ready to Land?
[ ] Build is completely green
Submissions with test failures require tracking issue and approval of a CODEOWNER
[ ] At least one +1 review by a CODEOWNER
[ ] All -1 reviews are confirmed resolved by the reviewer
Override/Marking reviews stale must be discussed with CODEOWNERS first
We might need to make exceptions for this in GA'd types as its a breaking change. Technically if someone is using that ctor today it wouldn't work so it might be an acceptable break I have asked @KrzysztofCwalina to weigh in before we move forward.
| gharchive/pull-request | 2022-08-01T05:20:41 | 2025-04-01T06:36:44.678427 | {
"authors": [
"Yao725",
"m-nash"
],
"repo": "Azure/autorest.csharp",
"url": "https://github.com/Azure/autorest.csharp/pull/2520",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1568110280 | Reduce Number of Methods Generates in Clients
Currently, depending on the configuration setup, in ClientMethodMapper we can end up generating overloads of methods that'll never be used. For example, if Context is enabled we'll end up with api(params) and api(paramsPlusContext) and if this is being used with a hand-written SDK the non-Context overload will almost never be called as the written layer will handle either propagating a Context or use an empty Context. This has a few downsides, it increases the size of JARs, results in longer compilation times, and skews test code coverage, given that we should look into the following ideas/improvements:
Only generate a Context overload if the generated client will be in implementation. The service client being in implementation is a good indicator on whether the generated code is being used in a hand-written SDK.
Only generate Response-based APIs if the generated client will be in implementation. The hand-written layer can handle cracking open the Response to return T.
If the client is generated only for implementation, one might actually use the sync-methods=NONE to only generate the Response methods (that likely be the minimum), and wrapper client can do whatever they want?
| gharchive/issue | 2023-02-02T13:57:56 | 2025-04-01T06:36:44.681663 | {
"authors": [
"alzimmermsft",
"weidongxu-microsoft"
],
"repo": "Azure/autorest.java",
"url": "https://github.com/Azure/autorest.java/issues/1942",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
447954117 | Parameter naming and ParameterSet issues in Update/Set cmdlets
In most Set and update cmdlets in generated compute code, There is an InputObjectParameter with Type IComputeIdentity and there is a separate Parameter Parameter which is the body parameter for the underlying object
Set cmdlets
(For example Set-AzVmss has a parameter called Parameter of type IVirtualMachineScaleSet`
This is in parameter sets containing both RGName + Name, and the InputObject parameter set.
[Parameter(ParameterSetName='Update1', Mandatory, HelpMessage='The name of the resource group.')]
[Parameter(ParameterSetName='UpdateExpanded1', Mandatory, HelpMessage='The name of the resource group.')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Category('Path')]
[System.String]
${ResourceGroupName},
[Parameter(ParameterSetName='Update1', Mandatory, HelpMessage='Subscription credentials which uniquely identify Microsoft Azure subscription. The subscription ID forms part of the URI for every service call.')]
[Parameter(ParameterSetName='UpdateExpanded1', Mandatory, HelpMessage='Subscription credentials which uniquely identify Microsoft Azure subscription. The subscription ID forms part of the URI for every service call.')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Category('Path')]
[System.String]
${SubscriptionId},
[Parameter(ParameterSetName='Update1', Mandatory, HelpMessage='The name of the VM scale set to create or update.')]
[Parameter(ParameterSetName='UpdateExpanded1', Mandatory, HelpMessage='The name of the VM scale set to create or update.')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Category('Path')]
[System.String]
${VMScaleSetName},
[Parameter(ParameterSetName='UpdateViaIdentity1', Mandatory, ValueFromPipeline, HelpMessage='Identity Parameter')]
[Parameter(ParameterSetName='UpdateViaIdentityExpanded1', Mandatory, ValueFromPipeline, HelpMessage='Identity Parameter')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Category('Path')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Models.IComputeIdentity]
${InputObject},
[Parameter(ParameterSetName='Update1', ValueFromPipeline, HelpMessage='Describes a Virtual Machine Scale Set.')]
[Parameter(ParameterSetName='UpdateViaIdentity1', ValueFromPipeline, HelpMessage='Describes a Virtual Machine Scale Set.')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Category('Body')]
[Microsoft.Azure.PowerShell.Cmdlets.Compute.Models.Api20190301.IVirtualMachineScaleSet]
${Parameter},
There are a few problems:
Both InputObject and Parameter both have 'ValueFromPipeline' set int he same parameter set, which is not possible to bind
Parameter is the name of the body parameter (instead of InputObject
I think the parameter set should be:
Path Parameters + Flattened body parameters
IdentityObject + Flattened body parameters
Where, in the second parameter set, if the object is a Vmss, we override its properties with the other parameters that are set. Or, thinking about it another way, we use this object as the body parameter and substitute the provided parameters
Update cmdlets
For Update (PATCH)
The same pattern occurs. I think the proper pattern is:
Path parameters + Optional flattened body parameters
ComputeIdentityObject (from pipeline) + optional flattened body parameters
In this case, the semantics should be to always use the input object only for its identity (and potentially, ETags)
This is occuriing in the latest compuet spec (examples from Set/Updat-Vmss): https://github.com/Azure/azure-powershell/blob/generation/src/Compute/readme.md
@fearthecowboy I think tha, for Update cmdlets we have a wowrking model, and perhaps for New as well.
For set cmdlets, the simplest solution may be to remove ViaIdentity parameter sets, and ensure that the type converter allows piping to the non-expanded parameter sets.
Short term:
remove non expanded variants
Longer:
have a varaiant that is like via identity but gets path from values in object
For now, we will just remove the non-expanded parameter sets for Set
Closing; no more immediate work
| gharchive/issue | 2019-05-24T02:43:40 | 2025-04-01T06:36:44.689204 | {
"authors": [
"fearthecowboy",
"markcowl"
],
"repo": "Azure/autorest.powershell",
"url": "https://github.com/Azure/autorest.powershell/issues/308",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
280707394 | global configuration files created with aztk spark init --global are not found in other dirs
When I run:
aztk spark init --global
a .aztk/ directory is correctly created in my $HOME dir. I edit the secrets.yaml file to configure keys, account names etc. But when I try and create a cluster in another location using the command:
aztk spark cluster create --id spark --vm-size standard_a2 --size 4
I get the message:
Batch account key is not set in secrets.yaml config
The workaround is to copy the directory ./aztk/ in my $HOME dir to the location I'm trying to create the cluster. I'm using the latest release of aztk with python 3.6.2 on MacOS High Sierra 10.13.1
conda list gives:
azure-batch 3.0.0
azure-common 1.1.8
azure-nspkg 2.0.0
azure-storage 0.33.0
I believe the wrong file path was being used to read the global secrets file. Just put out a PR to fix it. Nice catch!
| gharchive/issue | 2017-12-09T10:17:14 | 2025-04-01T06:36:44.698919 | {
"authors": [
"jafreck",
"willferreira"
],
"repo": "Azure/aztk",
"url": "https://github.com/Azure/aztk/issues/260",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1025442953 | bump aks-preview to 0.5.35 for mig support
This checklist is used to make sure that common guidelines for a pull request are followed.
General Guidelines
[ ] Have you run azdev style <YOUR_EXT> locally? (pip install azdev required)
[ ] Have you run python scripts/ci/test_index.py -q locally?
For new extensions:
[ ] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your PR is merged into master branch, a new PR will be created to update src/index.json automatically.
The precondition is to put your code inside this repo and upgrade the version in the PR but do not modify src/index.json.
forgot to do this in #3895
aks-preview
| gharchive/pull-request | 2021-10-13T16:12:12 | 2025-04-01T06:36:44.702430 | {
"authors": [
"alexeldeib",
"yonzhan"
],
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/3958",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1378621466 | [ssh] Bug Fix: Change how ssh vm reads the ssh client logs
Fix: https://github.com/Azure/azure-cli-extensions/issues/4915#issuecomment-1243939019
ssh vm reads the logs from ssh client and ssh proxy to delete sensitive information when it is no longer needed and print helpful error messages for known errors. Our approach to do that was causing ssh banners to be printed after the ssh process was terminated, and not before authentication.
Known issue: we are no longer able to read ssh proxy error channel when the user is running the command from a linux machine to connect to a local user on an arc machine. That is not a breaking change, but users in this particular situation will not get a helpful error message if there is an error with the proxy.
This checklist is used to make sure that common guidelines for a pull request are followed.
Related command
General Guidelines
[x] Have you run azdev style <YOUR_EXT> locally? (pip install azdev required)
[x] Have you run python scripts/ci/test_index.py -q locally?
For new extensions:
[x] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
The precondition is to put your code inside this repository and upgrade the version in the pull request but do not modify src/index.json.
ssh
| gharchive/pull-request | 2022-09-19T23:07:06 | 2025-04-01T06:36:44.707193 | {
"authors": [
"vthiebaut10",
"yonzhan"
],
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/5366",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1651387103 | {AzureAppService} Fixing the supported api-version for get_kube_client() function
Fixes Azure/azure-cli-extensions#6155
We get the below error while running CLI command: az appservice kube show
Enviornment:
Windows-10-10.0.20348-SP0
Python 3.10.10
Installer: MSI
azure-cli 2.47.0
Extensions:
appservice-kube 0.1.7
Error:
File "C:\Users\Administrator\.azure\cliextensions\appservice-kube\azext_appservice_kube\custom.py", line 305, in show_kube_environments
client = _get_kube_client(cmd)
File "C:\Users\Administrator\.azure\cliextensions\appservice-kube\azext_appservice_kube\custom.py", line 301, in _get_kube_client
return client.kube_environments
File "D:\a\_work\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/mgmt/web/_web_site_management_client.py", line 627, in kube_environments
ModuleNotFoundError: No module named 'azure.mgmt.web.v2021_01_01'
This error is because API version 2021-01-01 is trimmed from azure-mgmt-web SDK by scripts/trim_sdk.py to reduce MSI package size (#23946)
but appservice-kube extension hardcodes 2021-01-01:
https://github.com/Azure/azure-cli-extensions/blob/f4033ee2830c68668f340e12c39d76b6d2b7ee3f/src/appservice-kube/azext_appservice_kube/custom.py#L300
appservice-kube extension can use the latest API version api_version="2022-03-01" defined at:
https://github.com/Azure/azure-cli/blob/8c45b0feeaaa9a8ca3ed53c5cead1a8435b04fa6/src/azure-cli-core/azure/cli/core/profiles/_shared.py#L243
This checklist is used to make sure that common guidelines for a pull request are followed.
Related command
General Guidelines
[ ] Have you run azdev style <YOUR_EXT> locally? (pip install azdev required)
[ ] Have you run python scripts/ci/test_index.py -q locally?
For new extensions:
[ ] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
You only need to update the version information in file setup.py and historical information in file HISTORY.rst in your PR but do not modify src/index.json.
AzureAppService
@StrawnSC Could you please review this once you get a chance ? Thanks in advance.
ERROR cli.azure.cli.core.azclierror:azlogging.py:212 Can't overwrite existing cassette ('/mnt/vss/_work/1/s/src/appservice-kube/azext_appservice_kube/tests/latest/recordings/test_win_webapp_quick_create_runtime_kube.yaml') in your current record mode ('once').
No match for the request (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/clitest000001?api-version=2022-09-01>) was found.
Found 2 similar requests with 1 different matcher(s) :
1 - (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/clitest000001?api-version=2021-04-01>)..)
Matchers succeeded : ['method', 'scheme', 'host', 'port', 'path']
Matchers failed :
_custom_request_query_matcher - assertion failure :
None
The CI issues are caused by the bumping api-version of Resource RP from 2021-04-01 to 2022-09-01 in azure-cli-core, so you need to pull the latest code from the remote dev branch of Azure/azure-cli repo and then re-record these failed tests in live mode.
| gharchive/pull-request | 2023-04-03T05:58:49 | 2025-04-01T06:36:44.718264 | {
"authors": [
"navba-MSFT",
"yonzhan",
"zhoxing-ms"
],
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/6157",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1897597343 | Upgrade az vmware extension to use 2023-03-01 API version
az vmware: update to AVS 2023-03-01 API
az vmware private-cloud create: Add new --ext-nw-blocks parameter
az vmware private-cloud update: Add new --ext-nw-blocks parameter
This checklist is used to make sure that common guidelines for a pull request are followed.
Related command
General Guidelines
[ ] Have you run azdev style <YOUR_EXT> locally? (pip install azdev required)
[ ] Have you run python scripts/ci/test_index.py -q locally?
For new extensions:
[ ] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
You only need to update the version information in file setup.py and historical information in file HISTORY.rst in your PR but do not modify src/index.json.
Thank you for your contribution! We will review the pull request and get back to you soon.
Hi @jonathanhe-msft and @cataggar. I've changed the related the code. Could you help update the test recordings in your mock servers. Thanks.
az vmware placement-policy vm-host and az vmware placement-policy vm aren't working properly as they're missing a required affinityType parameter: https://github.com/Azure/azure-rest-api-specs/blob/main/specification/vmware/resource-manager/Microsoft.AVS/stable/2023-03-01/vmware.json#L7718
az vmware hcx-enterprise-site-create failing due to bad request. the request looks fine to me, but the mock server isn't accepting the values. I think it's an error with the mock, but can you confirm everything looks ok from your end?
vmware private-cloud add-identity-source fails with ResourceNotFoundError, even though i see the calls made successfully to our mock server:
aren't working properly as they
jonathanhe-msft Could you use the --debug argument to check the api response. It seems the affinityType isn't returned in the api response of Get method. The update command will GET the resource first. And as affinityType is not exposed as an argument, the value will be inherent from the GET response. According to api specs, that property should be included in api response.
az vmware hcx-enterprise-site-create failing due to bad request. the request looks fine to me, but the mock server isn't accepting the values. I verified with the old implementation (pre-migration) and it is working as expected. can you take a look?
Fixed in this commit(https://github.com/Azure/azure-cli-extensions/pull/6765/commits/7d7faac452443a57c953f1806df6300074db8cd3)
vmware private-cloud add-identity-source fails with ResourceNotFoundError, even though i see the calls made successfully to our mock server:
That's because the
vmware private-cloud add-identity-source fails with ResourceNotFoundError, even though i see the calls made successfully to our mock server:
That's because the newly added identity source is missed in the response of your mock server. I also find this issue in old recordings and I manually fixed the server response in previous migration PR.
az vmware placement-policy vm-host and az vmware placement-policy vm aren't working properly as they're missing a required affinityType parameter: https://github.com/Azure/azure-rest-api-specs/blob/main/specification/vmware/resource-manager/Microsoft.AVS/stable/2023-03-01/vmware.json#L7718
jonathanhe-msft Could you use the --debug argument to check the api response. It seems the affinityType isn't returned in the api response of Get method. The update command will GET the resource first. And as affinityType is not exposed as an argument, the value will be inherent from the GET response. According to api specs, that property should be included in api response.
confirmed get response is missing the affinityType property. working on fixing and will update live tests once complete.
@zhoxing-ms Please help review this PR.
| gharchive/pull-request | 2023-09-15T02:30:40 | 2025-04-01T06:36:44.735539 | {
"authors": [
"jonathanhe-msft",
"kairu-ms",
"yonzhan"
],
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/6765",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1950838527 | Update API for Aks Edge Essentials
This PR updates the get calls for annotations to provide distribution names for aks-edge-k3s and aks-edge-k8s.
This checklist is used to make sure that common guidelines for a pull request are followed.
Related command
General Guidelines
[x] Have you run azdev style <YOUR_EXT> locally? (pip install azdev required)
[x] Have you run python scripts/ci/test_index.py -q locally? (pip install wheel==0.30.0 required)
For new extensions:
[x] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
You only need to update the version information in file setup.py and historical information in file HISTORY.rst in your PR but do not modify src/index.json.
Draft
Please add test for this new feature.
Please add test for this new feature.
When you do kubectl get nodes on that infrastructure, you get an api response. In the infrastructure, the distro name is aks-edge-k3s. The test is dependent on the underlying infrastructure. We worked with the customer on validating this feature on that particular infrastructure. It passed the validation for the aks-edge-essentials scenario where this feature was tested on on-prem server with aks-edge-k3s/k8s configuration.
| gharchive/pull-request | 2023-10-18T22:55:24 | 2025-04-01T06:36:44.740967 | {
"authors": [
"9lash",
"yanzhudd",
"yonzhan"
],
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/6873",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
541131612 | az webapp: Magic .azure files created in my development environment during deploy
az feedback auto-generates most of the information requested below, as of CLI version 2.0.62
Describe the bug
During the deployment two magic files were created in my development environment — .azure and .deployment. These files didn't have comments or links to any documentation regarding what these files were for, what they contained, or if the options were configurable
To Reproduce
Steps to reproduce the behavior.
Expected behavior
A clear and concise description of what you expected to happen.
Environment summary
Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (az --version) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows)
Additional context
Add any other context about the problem here.
Running az webapp up configures local context for which is included in the .Azure files see https://docs.microsoft.com/en-us/cli/azure/reference-index?view=azure-cli-latest#az-configure for more details.
Adding @btardif as FYI.
We should add a comment to the content of the files with links to more info or explanation of what they are and what they are used for... re-opening this and assigning to me to come up with plan
move to S166.
@btardif any update for this issue?
move to S167
@yonzhan , @btardif and I just had a quick chat about this issue and the webapp team is suspecting that the magical .azure file is being created via the CLI rather than the az webapp up command. There's confusion on what this file is and how to make use of it. It'd be great if we could investigate and confirm on what's happening and the next steps
@Juliehzl please help to investigate the issue mentioned above.
move to S169
@chenlomis this is the same issue being discussed here: https://dev.azure.com/azuretracking/Azure tracking/_workitems/edit/680
add to S173
.azure should be a folder instead of a file, created by Azure CLI as the configuration folder.
https://github.com/Azure/azure-cli/blob/3b4d10d473a8d72dc2a1607478beb14424ce2004/src/azure-cli-core/azure/cli/core/_environment.py#L12
It is by-default created under the user's home folder, unless
AZURE_EXTENSION_DIR is set
az configure --scope local is used
local-context is turn on
.deployment seems to be created by botservice:
https://github.com/Azure/azure-cli/blob/9370ec21d5febcb0e7099f1a3969a42d04c4ceb1/src/azure-cli/azure/cli/command_modules/botservice/custom.py#L560
Please share the full deployment script for us to further analyze the issue.
The deployment path used is here:
https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-python?tabs=bash
This generates a .azure/config file with no comment explaining what it is for.
The .deployment file comes from VS Code deploy so that is a separate issue.
We have a document Azure CLI configuration explaining what .azure/config is.
Indeed, we can put a comments in .azure/config explaining what it is. @yonzhan @achandmsft @chenlomis, any thoughts?
Sounds like a plan to me
Something along the lines of:" this is where your CLI configs are stored"
@qubitron that aligns with your proposal right?
Yep something like "This file contains saved configurations for your CLI, read more at: aka.ms/1234".
We probably want an aka link in case we change where this points to.
@jiasli @chenlomis should we re-open the issue?
Created issue https://github.com/Azure/azure-cli/issues/15001 to track.
| gharchive/issue | 2019-12-20T18:19:11 | 2025-04-01T06:36:44.754809 | {
"authors": [
"btardif",
"chenlomis",
"jiasli",
"panchagnula",
"qubitron",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/11634",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
556526714 | Very ugly error message when using az account logging update command on a Storage Account of kind BlobStorage or sku.tier Premium
If you try to exeute az account logging update with a Premium storage account or with a BlobStorage storage account, you will get an error similar to the following:
az>> az storage logging show --account-name xxxxxxxxxxx --subscription xx-xxx-xxx
The command failed with an unexpected error. Here is the traceback:
HTTPSConnectionPool(host='xxxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connection.py", line 159, in _new_conn
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\util\connection.py", line 57, in create_connection
File "C:\Program Files (x86)\Microsoft SDKs\Azure\CLI2\lib\socket.py", line 745, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno 11002] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 600, in urlopen
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 343, in _make_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 839, in _validate_conn
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connection.py", line 301, in connect
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connection.py", line 168, in _new_conn
urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\adapters.py", line 449, in send
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 638, in urlopen
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\util\retry.py", line 399, in increment
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='xxxxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common\storageclient.py", line 321, in _perform_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common_http\httpclient.py", line 92, in perform_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\sessions.py", line 533, in request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\sessions.py", line 646, in send
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\adapters.py", line 516, in send
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='xxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\bhulse.azure\cliextensions\interactive\azext_interactive\azclishell\app.py", line 662, in cli_execute
result = invocation.execute(args)
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init_.py", line 603, in execute
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init_.py", line 661, in run_jobs_serially
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init.py", line 652, in run_job
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands\arm.py", line 750, in show_exception_handler
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init.py", line 631, in run_job
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init.py", line 306, in call
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core_init_.py", line 485, in default_command_handler
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli\azure\cli\command_modules\storage\operations\logging.py", line 15, in get_logging
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli\azure\cli\command_modules\storage\services_wrapper.py", line 29, in get_logging
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\queue\queueservice.py", line 372, in get_queue_service_properties
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common\storageclient.py", line 430, in _perform_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common\storageclient.py", line 361, in _perform_request
azure.common.AzureException: HTTPSConnectionPool(host='xxxxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
I would have expected a simple error message along the lines of:
Premium Storage accounts do not support logging
BlobStorage storage accounts do not support logging.
I though I was having a problem with api quotas.
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: c237378a-a708-8e2f-3c8c-0bc8b623d21c
Version Independent ID: c4e27b67-9605-60f5-899d-8ffd727079b2
Content: az storage logging
Content Source: src/azure-cli/azure/cli/command_modules/storage/_help.py
Service: storage
GitHub Login: @rloutlaw
Microsoft Alias: routlaw
add to S168
Hi @hulseware, sorry for late response first.
There are several points to clarify here:
Are you using az interactive to run cli commands?
When I try to reproduce your issue with normal cli (not in interactive mode), I only get one exception not four exceptions like what in your scenario. Here is what I get:
HTTPSConnectionPool(host='zuhblob.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.HTTPSConnect
ion object at 0x000001F15CE28288>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
Your suggestion is very valuable to us. As you mentioned, the error occurs because of specific storage account type. Actually it is caused by default value (bqt) in --services parameter:
BlobStorage doesn't support logging for queue service;
PremiumStorage doesn't support logging for blob/queue/table services.
I will refine the error message to make it point to wrong storage account type.
Yes, I use az interactive while developing the right cli commands to add to utilities targeted at maintaining better security across the subscriptions I am managing for my team. So the extra tracebacks were coming from the az interactive tool.
Yes, I believe that message is what I see when running via the normal command line.
Thanks for working to make the error message more descriptive of the real problem!
From: Zunli Hu notifications@github.com
Sent: Thursday, May 21, 2020 2:11 AM
To: Azure/azure-cli azure-cli@noreply.github.com
Cc: Hulse, Bruce bhulse@ptc.com; Mention mention@noreply.github.com
Subject: Re: [Azure/azure-cli] Very ugly error message when using az account logging update command on a Storage Account of kind BlobStorage or sku.tier Premium (#11969)
External email from: noreply@github.commailto:noreply@github.com
Hi @hulsewarehttps://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fhulseware&data=02|01|bhulse%40ptc.com|f25ae0dc795942b8432f08d7fd4dd671|b9921086ff774d0d828acb3381f678e2|0|0|637256383063474220&sdata=Zo%2FTrRWewqd33HZtjCS%2FtKAUGF%2FCs%2F2qoIMW0OlqI28%3D&reserved=0, sorry for late response first.
There are several points to clarify here:
Are you using az interactive to run cli commands?
When I try to reproduce your issue with normal cli (not in interactive mode), I only get one exception not four exceptions like what in your scenario. Here is what I get:
HTTPSConnectionPool(host='zuhblob.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.HTTPSConnect
ion object at 0x000001F15CE28288>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
Your suggestion is very valuable to us. As you mentioned, the error occurs because of specific storage account type. Actually it is caused by default value (bqt) in --services parameter:
BlobStorage doesn't support logging for queue service;
PremiumStorage doesn't support logging for blob/queue/table services.
I will refine the error message to make it point to wrong storage account type.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHubhttps://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2FAzure%2Fazure-cli%2Fissues%2F11969%23issuecomment-631902964&data=02|01|bhulse%40ptc.com|f25ae0dc795942b8432f08d7fd4dd671|b9921086ff774d0d828acb3381f678e2|0|0|637256383063479216&sdata=HCtaDxvR6agdww5TSxgEOE6WmNgJ4eTYKG3mk9%2Bpw00%3D&reserved=0, or unsubscribehttps://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FACLZFQQTT6JLU27D6HU6GFLRSTAYXANCNFSM4KM3LHYQ&data=02|01|bhulse%40ptc.com|f25ae0dc795942b8432f08d7fd4dd671|b9921086ff774d0d828acb3381f678e2|0|0|637256383063484198&sdata=Dj9dgmYAByp4XMcbQz9VIO6FfXbVtopi33iJZVX1u9A%3D&reserved=0.
| gharchive/issue | 2020-01-28T22:46:53 | 2025-04-01T06:36:44.791317 | {
"authors": [
"Juliehzl",
"hulseware",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/11969",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
624033610 | az cli fails with ModuleNotFoundError
This is autogenerated. Please review and update as needed.
Describe the bug
Command Name
az webapp create
Errors:
No module named 'decorator'
Traceback (most recent call last):
python3/dist-packages/fabric/connection.py, ln 5, in <module>
from invoke.vendor.six import StringIO
ModuleNotFoundError: No module named 'invoke.vendor.six'
...
python3/dist-packages/fabric/connection.py, ln 10, in <module>
from decorator import decorator
ModuleNotFoundError: No module named 'decorator'
To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
Install Ubuntu 20.4 LTS on Windows WSL
Install az cli following these steps: https://docs.microsoft.com/en-us/cli/azure/install-azure-cli-apt?view=azure-cli-latest#set-release
az login
az webapp create --name --resource-group --plan --runtime {}
Expected Behavior
az commands work
Environment Summary
Linux-4.4.0-18362-Microsoft-x86_64-with-glibc2.29
Python 3.8.2
Shell: bash
azure-cli 2.0.81
Extensions:
azure-devops 0.17.0
Additional Context
webapp
No related to app service - see https://github.com/Azure/azure-cli/issues/13644
I read all that at the time, I still couldn’t make it work on a fresh install following all the instructions. Care to elaborate? Someone else’s job?
@fengzhou-msft could you help with this?
@cuzzlor you need to first remove the packages installed by the old 2.0.81 azure-cli that comes with the Ubuntu 20.4 LTS. Try sudo apt remove azure-cli and make sure /usr/lib/python3/dist-packages/azure is deleted.
Ok. At the time it was a brand new install of WSL and the azure cli install script. I’ll try again maybe something has changed. Thanks.
You may need to run sudo apt autoremove after sudo apt remove azure-cli to remove the old azure packages that azure-cli 2.0.81 depends on.
Duplicate with #14011
| gharchive/issue | 2020-05-25T04:16:32 | 2025-04-01T06:36:44.799367 | {
"authors": [
"cuzzlor",
"fengzhou-msft",
"panchagnula",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/13660",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
688434804 | How to pass an argument parameter without defining a value?
I am trying to achieve three scenarios for a command. Is there a way to do this ?
az postgres server create (no argument parameter passed)
az postgres server create --public-access (no value passed for argument parameter --public-access)
az postgres server create --public access ( a valid value is passed to argument parameter --public-access).
Is there a way to do (2) that is pass a parameter without a value ? I am running into the below error :
az postgres server create: error: argument --public-access: expected one argument.
If this is do-able, how should I differentiate between (1) and (2) that is whether the parameter is completely missing or is present with a missing value ?
postgres
@arde0708 what are the sample values of --public-access? What is the behavior for 2?
We have get_three_state_flag arg_type that can achieve 2 and 3, but it only support boolean values.
We would want to define three kind of values for --public-access
No value. Just the parameter is passed. Ex : 'az postgres flexible-server create --public-access' ( Adds neither a Vnet, nor a firewall rule)
Single IP. Ex : 'az postgres flexible-server create --public-access 10.23.202.44' (Behavior : Adds single firewall rule)
Multiple IP. Ex : 'az postgres flexible-server create --public-access 10.23.202.44-10.23.202.48' (Behavior : Adds range of firewall rules)
Don't pass the parameter itself 'Ex : az postgres flexible-server create' (Behavior : Adds Vnet, Subnet)
@yonzhan Single Server is being retired. We won't be fixing this at this point in time. Please, close this issue. Thanks!
| gharchive/issue | 2020-08-29T01:49:55 | 2025-04-01T06:36:44.806509 | {
"authors": [
"arde0708",
"fengzhou-msft",
"nachoalonsoportillo",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/14968",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1282938468 | [Stream Analytics] Bugs related to clusterid property
az feedback auto-generates most of the information requested below, as of CLI version 2.0.62
Related command
az stream-analytics job update
In our streaming job definition, we have the following nested cluster and id properties:
"properties": {
"cluster": {
"id": "subscriptions/{mysub}/resourceGroup/{myrg}/etc..."
}
}
Describe the bug
In the CLI, this cluster id property gets translated into just id:
o az stream-analytics job update --job-name 'asa-fleide-large396' --resource-group $rg --id $clusterId
We think this is a possible source of confusion for the user -- because this command is an operation on the job, it's reasonable to infer that id is the job id, not the cluster id. Could we possible change this parameter to clusterid?
Another possible scenario is that the user should be able to set this id value to null. I tried to do this via CLI (setting the value to null or empty string), but got the following error:
(LinkedInvalidPropertyId) Property id '' at path 'properties.cluster.id' is invalid. Expect fully qualified resource Id that start with '/subscriptions/{subscriptionId}' or '/providers/{resourceProviderNamespace}/'.
Code: LinkedInvalidPropertyId
Message: Property id '' at path 'properties.cluster.id' is invalid. Expect fully qualified resource Id that start with '/subscriptions/{subscriptionId}' or '/providers/{resourceProviderNamespace}/'.
Is there another way to set it this property to null? If not, can we remove this validation?
To Reproduce
az stream-analytics job update --job-name 'asa-fleide-large396' --resource-group $rg --id $clusterId
az stream-analytics job update --job-name 'asa-fleide-large396' --resource-group $rg --id ""
Expected behavior
property name is clusterid instead of id
user able to set id value to null
Environment summary
Additional context
route to CXP team
Noncustomer-reported issue. Adding Service team to look in this.
| gharchive/issue | 2022-06-23T21:03:12 | 2025-04-01T06:36:44.813247 | {
"authors": [
"navba-MSFT",
"roslynlu",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/22990",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1582443192 | az webapp up does re-create existing appservice pan and does not upload published .net app
az feedback auto-generates most of the information requested below, as of CLI version 2.0.62
Related command
az webapp up -n $app -g $grp -p $plan --sku F1 -l $loc -r "DOTNET|6.0"
Describe the bug
Azure cli 2.45.0 az webapp up does re-create existing appservice pan insted of using it and does not upload published .net app - instead the raw source files are uploaded
To Reproduce
env=$RANDOM
grp=appservices-$env
loc=westeurope
plan=appservices-$env
app=cli-api-$env
az group create -n $grp -l $loc
az appservice plan create -n $plan -g $grp --sku S1
az webapp create -n $app -g $grp --plan $plan --runtime "DOTNET|6.0"
dotnet new webapi -n cli-api --framework net6.0
cd cli-api
az webapp up -n $app -g $grp -p $plan --sku F1 -l $loc -r "DOTNET|6.0"
cd ..
Expected behavior
the existing app service plan should be used and the app that has been processed by dotnet publish should have been uploaded
Environment summary
Windows 11, WSL, Azure cli 2.45.0
Additional context
route to CXP team
Adding Service team to look into this.
Unable to repro. Closing for now as this has been open for a while with no recent activity. Please update to the latest CLI version and let us know if this is still an issue.
Also, when running az webapp up after already creating the app using az webapp create, you don't need to specify the runtime/location/SKU again in az webapp up. The cli output when running the command might say "Creating AppServicePlan..." even though it already exists. We are aware of that bug and are addressing that separately. See https://github.com/Azure/azure-cli/issues/22767 for more details on that issue.
| gharchive/issue | 2023-02-13T14:21:25 | 2025-04-01T06:36:44.821956 | {
"authors": [
"arambazamba",
"navba-MSFT",
"seligj95",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/25443",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1875271598 | Implement slot parameter for az functionapp function show
Related command
az functionapp function show
Is your feature request related to a problem? Please describe.
I can't find a way to run az functionapp function show and target a deployment slot (besides the default/production one).
It apparently does not offer a --slot parameter: https://learn.microsoft.com/en-us/cli/azure/functionapp/function?view=azure-cli-latest#az-functionapp-function-show
Describe the solution you'd like
Please consider adding a --slot parameter to az functionapp function show
Describe alternatives you've considered
I tried a workaround via modifying --name to match the test slots name, but that does not seem to work
Additional context
This is the context in which I would use the new parameter: https://github.com/luis261/aztraphile/blob/reintroduce-slot-bound-http-invocations/utils/aztra-utils.ps1#L334
Thank you for opening this issue, we will look into it.
Hello @luis261 Thank you for your feedback. This is a feature request I just added to our backlog. In the meantime, you would just have to make direct API calls to https://management.azure.com/subscriptions/<subscriptionId>/resourceGroups/<resourceGroup>/providers/Microsoft.Web/sites/<functionAppName>/slots/<slotName>/functions/<functionName>?api-version=2022-03-01.
Hi @kamperiadis - glad to know it made the backlog! Thanks for keeping me in the loop. Also good to know the alternative via the direct API call (:
| gharchive/issue | 2023-08-31T10:47:06 | 2025-04-01T06:36:44.829107 | {
"authors": [
"kamperiadis",
"luis261",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/27297",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
249661833 | Add a VM to an AvailabilitySet after creation.
Description
Is there an azure-cli equivalent of this powershell cmdlet? Described here:
https://www.petri.com/understanding-azure-availability-sets
https://docs.microsoft.com/en-us/powershell/module/azure/set-azureavailabilityset?view=azuresmps-4.0.0`
The first link has an example like this:
Get-AzureVM -ServiceName contosodc1 -Name contosodc1 | Set-AzureAvailabilitySet -AvailabilitySetName contosodcs | Update-AzureVM
Environment summary
Install Method: nightly
CLI Version: azure-cli (2.0.12+1.dev20170808)
OS Version: ubuntu 16.04 x64
Shell Type: bash
Hi @nneul there is no dedicated command for this. I tried to accomplish this via generic update and had no luck. It seems you must specify the availability set when you create the VM (at least currently in 2.0) We have an open project to improve the adding of VMs to load balancers and application gateway pools.
I'll investigate some potential solutions for this next sprint and see if we can get something into the next release. Ideas would be having an availability-set add-vm command, or adding availability set as strongly typed parameter on vm update.
Moving this into the same project as improving LB/AG experience.
Anything new here?
The item is on our backlog but not planned for upcoming work. @yugangw-msft does this fit into your upcoming sprint plans?
@tjprescott, I will take a look in the next sprint
I have confirmed with Compute team that adding an existing VM to an existing av-set only works with RDFE(cloud service). For Compute RP, the only time you can add to an av-set is during VM create.
| gharchive/issue | 2017-08-11T14:59:03 | 2025-04-01T06:36:44.834652 | {
"authors": [
"nneul",
"superminiek",
"tjprescott",
"yugangw-msft"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/4202",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
359013990 | az webapp deployment source config-zip fails from jenkins pipeline
Describe the bug
We are using Jenkins integration with Azure. While trying to run az deploy webapp from jenkins pipeline it fails with the following error:
az webapp deployment source config-zip --debug --resource-group st-worker-test --name st-worker-test --slot stagedeploy --src /var/lib/jenkins/workspace/tore-backend-latest_develop@2/artifact/App.Store.Work_develop_549.zip
ERROR: Expecting value: line 1 column 1 (char 0)
Traceback (most recent call last):
File "/opt/az/lib/python3.6/site-packages/knack/cli.py", line 197, in invoke
cmd_result = self.invocation.execute(args)
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/commands/init.py", line 369, in execute
six.reraise(*sys.exc_info())
File "/opt/az/lib/python3.6/site-packages/six.py", line 693, in reraise
raise value
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/commands/init.py", line 343, in execute
result = cmd(params)
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/commands/init.py", line 182, in call
return self.handler(*args, **kwargs)
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/init.py", line 436, in default_command_handler
result = op(**command_args)
File "/opt/az/lib/python3.6/site-packages/azure/cli/command_modules/appservice/custom.py", line 188, in enable_zip_deploy
response = response.json()
File "/opt/az/lib/python3.6/site-packages/requests/models.py", line 896, in json
return complexjson.loads(self.text, **kwargs)
File "/opt/az/lib/python3.6/json/init.py", line 354, in loads
return _default_decoder.decode(s)
File "/opt/az/lib/python3.6/json/decoder.py", line 339, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/opt/az/lib/python3.6/json/decoder.py", line 357, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
To Reproduce
Set to Azure serviceprincipal credentials
use Azure cli commands for the below:
login to Azure with service principal
set to the azure subscription
run az webapp deployment source config-zip --debug --resource-group st-worker-test --name st-worker-test --slot stagedeploy --src /var/lib/jenkins/workspace/tore-backend-latest_develop@2/artifact/App.Store.Work_develop_549.zip
The command above fails with the error mentioned.
Expected behavior
It should deploy the zip file to Azure webapp sucessfully
Environment summary
Install Method : sudo apt-get / CLI version (azure-cli (2.0.45)) / Ubuntu 16.0.4 / Shell Type (bash)
Additional context
If i run the same command manually by logging in to the Jenkins slave it works successfully. We need to get this working from Jenkins pipeline and need help in fixing the issue.
Also seeing that the deployment hangs sometimes. Attached is the debug log. I really need help on this. Could someone please look into it.
azure-deploy-log.docx
@panchagnula
@panchagnula I think the right fix here is to make the timeout configurable with an optional flag. The current behavior is to be assumed as the default if no timeout is provided.
Same here using on a bash terminal under macos v10.14.1. When deploying the demo project under https://docs.microsoft.com/en-us/learn/modules/store-app-data-with-azure-blob-storage/7-blob-uploads-and-downloads it fails with a highly similar error:
az webapp deployment source config-zip --resource-group ${AZ_RES_GROUP} --name ${AZ_WEBAPP_NAME} --src ../site.zip
Expecting value: line 1 column 1 (char 0)
Traceback (most recent call last):
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/knack/cli.py", line 197, in invoke
cmd_result = self.invocation.execute(args)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/commands/__init__.py", line 369, in execute
six.reraise(*sys.exc_info())
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/six.py", line 693, in reraise
raise value
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/commands/__init__.py", line 343, in execute
result = cmd(params)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/commands/__init__.py", line 182, in __call__
return self.handler(*args, **kwargs)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/__init__.py", line 436, in default_command_handler
result = op(**command_args)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/command_modules/appservice/custom.py", line 262, in enable_zip_deploy
response = response.json()
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/requests/models.py", line 892, in json
return complexjson.loads(self.text, **kwargs)
File "/Users/brandl/anaconda3/lib/python3.6/json/__init__.py", line 354, in loads
return _default_decoder.decode(s)
File "/Users/brandl/anaconda3/lib/python3.6/json/decoder.py", line 339, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/Users/brandl/anaconda3/lib/python3.6/json/decoder.py", line 357, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
My zip file seems legit, and my CLI version is azure-cli (2.0.48)
@holgerbrandl - what is the python version that is being used here? is it possible for you to upgrade the python version & try? Do Ptyhon --version, this is the version i have installed "Python 3.7.0". thank you.
Since I had deleted the corresponding resource group already, I've created the complete demo setup again (group, storage account, wegapp) and tried to reproduce the error. However now it seems fine and deployment works as expected.
I'm not aware of any changes in my local setup. My python version has not changed since I have reported the ticket and is 3.6.3.
Not sure if this is compatible with the stacktrace, but it may have been also some kind of connection timeout?
@holgerbrandl , thanks for the update. with the new setup has the version of azure-cli changed, especially since we have a newer version available now & we have had some code changes to the zip deployment command, so there might be some bugfixes that would have infact fixed the problem you are seeing.
| gharchive/issue | 2018-09-11T11:59:50 | 2025-04-01T06:36:44.849438 | {
"authors": [
"holgerbrandl",
"mattchenderson",
"panchagnula",
"swm7",
"tjprescott"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/7282",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1367411444 | [Core] PREVIEW: Support Web Account Manager (WAM) login on Windows
Related command
az login
Description
Close #21201
Require https://github.com/AzureAD/microsoft-authentication-library-for-python/pull/415
Rework https://github.com/Azure/azure-cli/pull/22774
Testing Guide
On Windows, users can opt in WAM login using:
az config set core.allow_broker=true
Then run
az login
Instead of launching a web browser, an account selector will be shown to ask the user to select a Windows account:
History Notes
[Core] PREVIEW: Support Web Account Manager (WAM) login on Windows. To opt in, run az config set core.allow_broker=true
Additional information
WAM will be enabled by default in year 2023.
WAM support
| gharchive/pull-request | 2022-09-09T07:29:28 | 2025-04-01T06:36:44.854852 | {
"authors": [
"jiasli",
"yonzhan"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/pull/23828",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
212290280 | Add profile switching params and profile listing command
Modify ‘az cloud register’ and ‘az cloud update’ to include the ‘—profile’ parameter
Add ability to view all supported API versions without querying a cloud
note: With this change, switching profile will not change anything as they are just example profiles.
FYI @lmazuel For the _shared.py file change. This would contain the resource type to api version map.
Closes https://github.com/Azure/azure-cli/issues/2280
Codecov Report
Merging #2398 into api-profile-support will increase coverage by <.01%.
The diff coverage is 77.14%.
@@ Coverage Diff @@
## api-profile-support #2398 +/- ##
=======================================================
+ Coverage 72.33% 72.33% +<.01%
=======================================================
Files 323 325 +2
Lines 18273 18304 +31
Branches 2701 2707 +6
=======================================================
+ Hits 13217 13240 +23
- Misses 4223 4231 +8
Partials 833 833
Impacted Files
Coverage Δ
...cli-cloud/azure/cli/command_modules/cloud/_help.py
100% <100%> (ø)
:white_check_mark:
src/azure-cli-core/azure/cli/core/cloud.py
69.08% <100%> (+2.58%)
:white_check_mark:
...-cloud/azure/cli/command_modules/cloud/commands.py
100% <100%> (ø)
:white_check_mark:
...azure-cli-core/azure/cli/core/profiles/init.py
100% <100%> (ø)
...i-cloud/azure/cli/command_modules/cloud/_params.py
93.33% <100%> (+1.02%)
:white_check_mark:
src/azure-cli-core/azure/cli/core/_profile.py
83.78% <100%> (-0.15%)
:x:
.../azure-cli-core/azure/cli/core/profiles/_shared.py
100% <100%> (ø)
...li-cloud/azure/cli/command_modules/cloud/custom.py
17.24% <11.11%> (-1.13%)
:x:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 4c0fdcf...52e3356. Read the comment docs.
| gharchive/pull-request | 2017-03-07T00:07:58 | 2025-04-01T06:36:44.869014 | {
"authors": [
"codecov-io",
"derekbekoe"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/pull/2398",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1475637951 | {Style} Use light theme by default on MacOS
Issue
We specify the color for warnings as ANSI color Bright Yellow without any specification of the actual RGB value:
https://github.com/Azure/azure-cli/blob/8d8d71707d6c60083d008062667cab18d0adaa25/src/azure-cli-core/azure/cli/core/style.py#L73
According to https://github.com/Azure/azure-cli/issues/18298#issuecomment-1263529261, MacOS's terminal by default uses white background and Bright Yellow is not readable on that white background.
Something like:
Personally, I think it should the terminal's theme's responsibility to make all ANSI colors readable on its background.
Change
Since MacOS's terminal by default uses white background, we by default uses light theme which uses (dark) Yellow for warnings:
https://github.com/Azure/azure-cli/blob/8d8d71707d6c60083d008062667cab18d0adaa25/src/azure-cli-core/azure/cli/core/style.py#L85
Cons
However, a disadvantage is that if the user manually chooses a dark background, or use an IDE like PyCharm with dark theme. There may be other readability issues if we default to light theme, such as blue, magenta.
As there isn't a way to detect the background color, there is no optimal solution. Users still need to run az config set core.theme=light/dark to select the color theme or totally disable color with az config set core.no_color=true.
https://stackoverflow.com/questions/58186391/text-displaying-as-bright-yellow-in-terminal
This is the interface for configuring ANSI color's real rendered RGB:
| gharchive/pull-request | 2022-12-05T04:29:12 | 2025-04-01T06:36:44.875038 | {
"authors": [
"jiasli"
],
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/pull/24804",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
306202432 | Failed to create File Store directory Error:mkdir : The system cannot find the path specified.
Running the new azure-vnet-plugin.exe 1.0.3 on my Windows 10 Azure VM gives me the below error...
C:\temp\azure-vnet-cnm-windows-amd64-v1.0.3>azure-vnet-plugin.exe
Failed to create File Store directory Error:mkdir : The system cannot find the path specified.
Also, am I supposed to set the --api-url argument? What value should go there?
Thanks,
J-F
Closing as this is now fixed in latest release.
| gharchive/issue | 2018-03-18T00:32:24 | 2025-04-01T06:36:44.876873 | {
"authors": [
"frenchyjef",
"sharmasushant"
],
"repo": "Azure/azure-container-networking",
"url": "https://github.com/Azure/azure-container-networking/issues/104",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
128376807 | Update resource-group-move-resources.md
Added Express Route as not support moving a resource
Hi @andyjballgit, I'm your friendly neighborhood Azure Pull Request Bot (You can call me AZPRBOT). Thanks for your contribution!
This seems like a small (but important) contribution, so no contribution license agreement is required at this point. Real humans will now evaluate your PR.
TTYL, AZPRBOT;
@ggailey777 - I have double-checked this change. This pull request in azure-content can be merged.
Thanks for helping out with the Azure docs
| gharchive/pull-request | 2016-01-24T06:28:05 | 2025-04-01T06:36:44.879232 | {
"authors": [
"andyjballgit",
"azurecla",
"ggailey777",
"tfitzmac"
],
"repo": "Azure/azure-content",
"url": "https://github.com/Azure/azure-content/pull/5442",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
160245572 | fixed minor typo
fixed minor typo
Hi @victorar, I'm your friendly neighborhood Azure Pull Request Bot (You can call me AZPRBOT). Thanks for your contribution!
It looks like you're working at Microsoft (victorar). If you're full-time, we DON'T require a contribution license agreement.
If you are a vendor, DO please sign the electronic contribution license agreement. It will take 2 minutes and there's no faxing! https://cla.azure.com.
TTYL, AZPRBOT;
@victorar : Thanks for your contribution to the Azure documentation! The author, @telmosampaio, has been notified to review your proposed change.
@telmosampaio
Easy typo. I'm accepting this PR. Thanks!
--Carolyn
@victorar
Nice catch! I'll accept this PR. Thank you!
--Carolyn
| gharchive/pull-request | 2016-06-14T18:02:24 | 2025-04-01T06:36:44.882896 | {
"authors": [
"PRmerger",
"azurecla",
"cjgronlund",
"victorar"
],
"repo": "Azure/azure-content",
"url": "https://github.com/Azure/azure-content/pull/6822",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1523359111 | Add pid to debug attach printout
Add pid to debug attach printout. Since attaching debugger to the process requires the PID, having it displayed saves an extra step and avoids confusion when multiple instances of azd is running.
Azure Dev CLI Install Instructions
Install scripts
MacOS/Linux
May elevate using sudo on some platforms and configurations
bash:
curl -fsSL https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/uninstall-azd.sh | bash;
curl -fsSL https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/install-azd.sh | bash -s -- --base-url https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329 --version '' --verbose
pwsh:
Invoke-RestMethod 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/uninstall-azd.ps1' -OutFile uninstall-azd.ps1; ./uninstall-azd.ps1
Invoke-RestMethod 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/install-azd.ps1' -OutFile install-azd.ps1; ./install-azd.ps1 -BaseUrl 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329' -Version '' -Verbose
Windows
PowerShell install
powershell -c "Set-ExecutionPolicy Bypass Process; irm 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/uninstall-azd.ps1' > uninstall-azd.ps1; ./uninstall-azd.ps1;"
powershell -c "Set-ExecutionPolicy Bypass Process; irm 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/install-azd.ps1' > install-azd.ps1; ./install-azd.ps1 -BaseUrl 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329' -Version '' -Verbose;"
MSI install
powershell -c "irm 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-windows-amd64.msi' -OutFile azd-windows-amd64.msi; msiexec /i azd-windows-amd64.msi /qn"
Standalone Binary
Linux - https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-linux-amd64.tar.gz
MacOS - https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-darwin-amd64.zip
Windows - https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-windows-amd64.zip
MSI
https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-windows-amd64.msi
Container
docker run -it azdevcliextacr.azurecr.io/azure-dev:pr-1329
Documentation
learn.microsoft.com documentation
title: Azure Developer CLI Preview reference
description: This article explains the syntax and parameters for the various Azure Developer CLI Preview commands.
author: puichan
ms.author: puichan
ms.date: 01/07/23
ms.topic: conceptual
ms.custom: devx-track-azdevcli
ms.prod: azure
Azure Developer CLI Preview reference
This article explains the syntax and parameters for the various Azure Developer CLI Preview commands.
azd
Azure Developer CLI is a command-line interface for developers who build Azure solutions.
Synopsis
Azure Developer CLI is a command-line interface for developers who build Azure solutions.
To begin working with Azure Developer CLI, run the azd up command by supplying a sample template in an empty directory:
azd up –-template todo-nodejs-mongo
You can pick a template by running azd template listand then supplying the repo name as a value to --template.
The most common next commands are:
azd pipeline config
azd deploy
azd monitor --overview
For more information, visit the Azure Developer CLI Dev Hub: https://aka.ms/azure-dev/devhub.
Options
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
-h, --help Gets help for azd.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
azd deploy: Deploy the application's code to Azure.
azd down: Delete Azure resources for an application.
azd env: Manage environments.
azd infra: Manage Azure resources.
azd init: Initialize a new application.
azd login: Log in to Azure.
azd logout: Log out of Azure
azd monitor: Monitor a deployed application.
azd pipeline: Manage GitHub Actions pipelines.
azd provision: Provision the Azure resources for an application.
azd restore: Restore application dependencies.
azd template: Manage templates.
azd up: Initialize application, provision Azure resources, and deploy your project with a single command.
azd version: Print the version number of Azure Developer CLI.
azd config
Manage Azure Developer CLI configuration
Synopsis
Manage the Azure Developer CLI user configuration, which includes your default Azure subscription and location.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
Options
-h, --help Gets help for config.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config get: Gets a configuration
azd config list: Lists all configuration values
azd config reset: Resets configuration to default
azd config set: Sets a configuration
azd config unset: Unsets a configuration
Back to top
azd config get
Gets a configuration
Synopsis
Gets a configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config get <path> [flags]
Options
-h, --help Gets help for get.
-o, --output string The output format (the supported formats are json). (default "json")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config list
Lists all configuration values
Synopsis
Lists all configuration values in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config list [flags]
Options
-h, --help Gets help for list.
-o, --output string The output format (the supported formats are json). (default "json")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config reset
Resets configuration to default
Synopsis
Resets all configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable to the default.
azd config reset [flags]
Options
-h, --help Gets help for reset.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config set
Sets a configuration
Synopsis
Sets a configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config set <path> <value> [flags]
Examples
azd config set defaults.subscription <yourSubscriptionID>
azd config set defaults.location eastus
Options
-h, --help Gets help for set.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config unset
Unsets a configuration
Synopsis
Removes a configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config unset <path> [flags]
Examples
azd config unset defaults.location
Options
-h, --help Gets help for unset.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd deploy
Deploy the application's code to Azure.
Synopsis
Deploy the application's code to Azure.
When no --service value is specified, all services in the azure.yaml file (found in the root of your project) are deployed.
Examples:
azd deploy
azd deploy --service api
azd deploy --service web
After the deployment is complete, the endpoint is printed. To start the service, select the endpoint or paste it in a browser.
azd deploy [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for deploy.
-o, --output string The output format (the supported formats are json, none). (default "none")
--service string Deploys a specific service (when the string is unspecified, all services that are listed in the azure.yaml file are deployed).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd down
Delete Azure resources for an application.
azd down [flags]
Options
-e, --environment string The name of the environment to use.
--force Does not require confirmation before it deletes resources.
-h, --help Gets help for down.
-o, --output string The output format (the supported formats are json, none). (default "none")
--purge Does not require confirmation before it permanently deletes resources that are soft-deleted by default (for example, key vaults).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd env
Manage environments.
Synopsis
Manage environments.
With this command group, you can create a new environment or get, set, and list your application environments. An application can have multiple environments (for example, dev, test, prod), each with a different configuration (that is, connectivity information) for accessing Azure resources.
You can find all environment configurations under the .azure<environment-name> folder. The environment name is stored as the AZURE_ENV_NAME environment variable in the .azure<environment-name>\folder.env file.
Options
-h, --help Gets help for env.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env get-values: Get all environment values.
azd env list: List environments
azd env new: Create a new environment.
azd env refresh: Refresh environment settings by using information from a previous infrastructure provision.
azd env select: Set the default environment.
azd env set: Set a value in the environment.
Back to top
azd env get-values
Get all environment values.
azd env get-values [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for get-values.
-o, --output string The output format (the supported formats are json, dotenv). (default "dotenv")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env list
List environments
azd env list [flags]
Options
-h, --help Gets help for list.
-o, --output string The output format (the supported formats are json, table). (default "table")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env new
Create a new environment.
azd env new <environment> [flags]
Options
-h, --help Gets help for new.
-l, --location string Azure location for the new environment
--subscription string Name or ID of an Azure subscription to use for the new environment
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env refresh
Refresh environment settings by using information from a previous infrastructure provision.
azd env refresh [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for refresh.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env select
Set the default environment.
azd env select <environment> [flags]
Options
-h, --help Gets help for select.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env set
Set a value in the environment.
azd env set <key> <value> [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for set.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd infra
Manage Azure resources.
Options
-h, --help Gets help for infra.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd infra create: Create Azure resources for an application.
azd infra delete: Delete Azure resources for an application.
Back to top
azd infra create
Create Azure resources for an application.
azd infra create [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for create.
--no-progress Suppresses progress information.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd infra: Manage Azure resources.
Back to top
azd infra delete
Delete Azure resources for an application.
azd infra delete [flags]
Options
-e, --environment string The name of the environment to use.
--force Does not require confirmation before it deletes resources.
-h, --help Gets help for delete.
--purge Does not require confirmation before it permanently deletes resources that are soft-deleted by default (for example, key vaults).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd infra: Manage Azure resources.
Back to top
azd init
Initialize a new application.
Synopsis
Initialize a new application.
When no template is supplied, you can optionally select an Azure Developer CLI template for cloning. Otherwise, azd init initializes the current directory and creates resources so that your project is compatible with Azure Developer CLI.
When a template is provided, the sample code is cloned to the current directory.
azd init [flags]
Options
-b, --branch string The template branch to initialize from.
-e, --environment string The name of the environment to use.
-h, --help Gets help for init.
-l, --location string Azure location for the new environment
--subscription string Name or ID of an Azure subscription to use for the new environment
-t, --template string The template to use when you initialize the project. You can use Full URI, <owner>/<repository>, or <repository> if it's part of the azure-samples organization.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd login
Log in to Azure.
Synopsis
Log in to Azure.
When run without any arguments, log in interactively using a browser. To log in using a device code, pass
--device-code.
To log in as a service principal, pass --client-id and --tenant-id as well as one of --client-secret,
--client-certificate, --client-credential or --client-credential-provider.
azd login [flags]
Options
--check-status Checks the log-in status instead of logging in.
--client-certificate string The path to the client certificate for the service principal to authenticate with.
--client-id string The client id for the service principal to authenticate with.
--client-secret string The client secret for the service principal to authenticate with. Set to the empty string to read the value from the console.
--federated-credential string The federated token for the service principal to authenticate with. Set to the empty string to read the value from the console.
--federated-credential-provider string The provider to use to acquire a federated token to authenticate with.
-h, --help Gets help for login.
-o, --output string The output format (the supported formats are json, none). (default "none")
--redirect-port int Choose the port to be used as part of the redirect URI during interactive login.
--tenant-id string The tenant id for the service principal to authenticate with.
--use-device-code When true, log in by using a device code instead of a browser.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd logout
Log out of Azure
Synopsis
Log out of Azure
azd logout [flags]
Options
-h, --help Gets help for logout.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd monitor
Monitor a deployed application.
Synopsis
Monitor a deployed application.
Examples:
azd monitor --overview
azd monitor -–live
azd monitor --logs
For more information, go to https://aka.ms/azure-dev/monitor.
azd monitor [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for monitor.
--live Open a browser to Application Insights Live Metrics. Live Metrics is currently not supported for Python applications.
--logs Open a browser to Application Insights Logs.
--overview Open a browser to Application Insights Overview Dashboard.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd pipeline
Manage GitHub Actions pipelines.
Synopsis
Manage GitHub Actions pipelines.
The Azure Developer CLI template includes a GitHub Actions pipeline configuration file (in the .github/workflows folder) that deploys your application whenever code is pushed to the main branch.
For more information, go to https://aka.ms/azure-dev/pipeline.
Options
-h, --help Gets help for pipeline.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd pipeline config: Create and configure your deployment pipeline by using GitHub Actions or Azure Pipelines.
Back to top
azd pipeline config
Create and configure your deployment pipeline by using GitHub Actions or Azure Pipelines.
Synopsis
Create and configure your deployment pipeline by using GitHub Actions or Azure Pipelines.
For more information, go to https://aka.ms/azure-dev/pipeline.
azd pipeline config [flags]
Options
--auth-type string The authentication type used between the pipeline provider and Azure for deployment (Only valid for GitHub provider)
-e, --environment string The name of the environment to use.
-h, --help Gets help for config.
--principal-name string The name of the service principal to use to grant access to Azure resources as part of the pipeline.
--principal-role string The role to assign to the service principal. (default "contributor")
--provider string The pipeline provider to use (github for Github Actions and azdo for Azure Pipelines). (default "github")
--remote-name string The name of the git remote to configure the pipeline to run on. (default "origin")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd pipeline: Manage GitHub Actions pipelines.
Back to top
azd provision
Provision the Azure resources for an application.
Synopsis
Provision the Azure resources for an application.
The command prompts you for the following:
Environment name: The name of your environment.
Azure location: The Azure location where your resources will be deployed.
Azure subscription: The Azure subscription where your resources will be deployed.
Depending on what Azure resources are created, running this command might take a while. To view progress, go to the Azure portal and search for the resource group that contains your environment name.
azd provision [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for provision.
--no-progress Suppresses progress information.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd restore
Restore application dependencies.
Synopsis
Restore application dependencies.
Run this command to download and install all the required libraries so that you can build, run, and debug the application locally.
For the best local run and debug experience, go to https://aka.ms/azure-dev/vscode to learn how to use the Visual Studio Code extension.
azd restore [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for restore.
--service string Restores a specific service (when the string is unspecified, all services that are listed in the azure.yaml file are restored).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd template
Manage templates.
Options
-h, --help Gets help for template.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd template list: List templates.
azd template show: Show the template details.
Back to top
azd template list
List templates.
azd template list [flags]
Options
-h, --help Gets help for list.
-o, --output string The output format (the supported formats are json, table). (default "table")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd template: Manage templates.
Back to top
azd template show
Show the template details.
azd template show <template> [flags]
Options
-h, --help Gets help for show.
-o, --output string The output format (the supported formats are json, table). (default "table")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd template: Manage templates.
Back to top
azd up
Initialize application, provision Azure resources, and deploy your project with a single command.
Synopsis
Initialize the project (if the project folder has not been initialized or cloned from a template), provision Azure resources, and deploy your project with a single command.
This command executes the following in one step:
azd init
azd provision
azd deploy
When no template is supplied, you can optionally select an Azure Developer CLI template for cloning. Otherwise, running azd up initializes the current directory so that your project is compatible with Azure Developer CLI.
azd up [flags]
Options
-b, --branch string The template branch to initialize from.
-e, --environment string The name of the environment to use.
-h, --help Gets help for up.
-l, --location string Azure location for the new environment
--no-progress Suppresses progress information.
-o, --output string The output format (the supported formats are json, none). (default "none")
--service string Deploys a specific service (when the string is unspecified, all services that are listed in the azure.yaml file are deployed).
--subscription string Name or ID of an Azure subscription to use for the new environment
-t, --template string The template to use when you initialize the project. You can use Full URI, <owner>/<repository>, or <repository> if it's part of the azure-samples organization.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd version
Print the version number of Azure Developer CLI.
azd version [flags]
Options
-h, --help Gets help for version.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
| gharchive/pull-request | 2023-01-07T00:45:22 | 2025-04-01T06:36:44.970964 | {
"authors": [
"azure-sdk",
"weikanglim"
],
"repo": "Azure/azure-dev",
"url": "https://github.com/Azure/azure-dev/pull/1329",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1351038098 | [HttpTrigger] Receiving 403 when not sending Expect: 100-continue
We are hosting an Azure Function V4 with runtime type dotnet-isolated.
The package versions are:
Microsoft.Azure.Functions.Worker "1.6.0"
Microsoft.Azure.Functions.Worker.Extensions.Http "3.0.13"
When the Function is deployed to Azure we are facing some issues with larger requests where the request body is over let's say 100 kb. We immediately get a 403 error. This doesn't even log a requests so we don't have any logging we can view. The only way we get it to work is by adding a Expect: 100-continue header. However we don't expect our users to always send this. How can we disable this feature?
@Ultre00 Is this a V3 app or V4? You can check the AzureFunctionsVersion element value in your .csproj file. Do you have a minimal repro you could share? (A github repo will be great)
@kshyju as mentioned this is a V4 function. I also figured out when this happens. It is only in combination with mTLS where the option "clientCertEnabled" is set to true. For now I simply set this to false since we are no longer using mTLS here.
@Ultre00 -- where did you have these configurations set? Is this somewhere in App Service?
@brettsam yes you can set that in the Function App. If you use an ARM template the properties are called
clientCertEnabled in combination with clientCertMode but if you use the UI then you can also do this under configuration > general settings > Client certificate mode
| gharchive/issue | 2022-08-25T14:57:47 | 2025-04-01T06:36:45.007411 | {
"authors": [
"Ultre00",
"brettsam",
"kshyju"
],
"repo": "Azure/azure-functions-dotnet-worker",
"url": "https://github.com/Azure/azure-functions-dotnet-worker/issues/1005",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
863938687 | Not able to debug AzF
Hi, I'm not able to debug. Once I attached written PID, I'm getting error on System.PrivateCore.Lib Can't write the message because the call is complete.
When I look inside, I'm gettting this:
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at Microsoft.Azure.Functions.Worker.GrpcWorker.d__14.MoveNext() in D:\a\1\s\src\DotNetWorker.Grpc\GrpcWorker.cs:line 86
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at Microsoft.Azure.Functions.Worker.WorkerHostedService.d__3.MoveNext() in D:\a\1\s\src\DotNetWorker.Core\WorkerHostedService.cs:line 27
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at Microsoft.Extensions.Hosting.Internal.Host.d__9.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.ConfiguredTaskAwaitable.ConfiguredTaskAwaiter.GetResult()
at Microsoft.Extensions.Hosting.HostingAbstractionsHostExtensions.d__4.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.Extensions.Hosting.HostingAbstractionsHostExtensions.d__4.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at xxx.Program.d__0.MoveNext() in
Do you know, what could be an issue?
@ondrejdurica can you share more details on how you're launching the worker? Is this from VS? Direct command line? If from VS, what does your project configuration look like for the debugging properties?
Thanks!
| gharchive/issue | 2021-04-21T14:13:59 | 2025-04-01T06:36:45.013400 | {
"authors": [
"fabiocav",
"ondrejdurica"
],
"repo": "Azure/azure-functions-dotnet-worker",
"url": "https://github.com/Azure/azure-functions-dotnet-worker/issues/406",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1264762713 | create_from_edge_environment does not send messages to IoT Hub
Hello,
I've been facing a weird issue using:
iotedge 1.2.10
edgeAgent/edgeHub 1.2.10.55800403
azure-iot-device 2.11.0
When I create the client module using create_from_edge_environment function within my IoT Edge Module, the module is able to correctly push and pull twin, but any message sent is never received on the hub.
If I use the from_connection_string and use a connection string within the same Docker, it works...
This is really strange, here some information I got from logs:
2022-06-08 09:13:18 azure.iot.device.common.mqtt_transport _create_mqtt_client 149 [INFO]: Creating client for connecting using MQTT over TCP
2022-06-08 09:13:18 azure.iot.device.common.http_transport request 108 [INFO]: sending https POST request to devices/Gateway-1200723-01004/files .
2022-06-08 09:13:19 azure.iot.device.iothub.aio.async_clients get_storage_info_for_blob 572 [INFO]: Successfully retrieved storage_info
2022-06-08 09:13:20 azure.iot.device.common.http_transport request 108 [INFO]: sending https POST request to devices/Gateway-1200723-01004/files/notifications .
2022-06-08 09:13:20 azure.iot.device.iothub.aio.async_clients notify_blob_upload_status 598 [INFO]: Successfully notified blob upload status
2022-06-08 09:13:20 azure.iot.device.iothub.aio.async_clients send_message 349 [INFO]: Sending message to Hub...
2022-06-08 09:13:20 azure.iot.device.common.mqtt_transport publish 528 [INFO]: publishing on devices/Gateway-1200723-01004/modules/GWApp/messages/events/
2022-06-08 09:13:20 azure.iot.device.common.mqtt_transport on_publish 243 [INFO]: payload published for 8
2022-06-08 09:13:20 azure.iot.device.iothub.aio.async_clients send_message 356 [INFO]: Successfully sent message to Hub
Jun 08 09:13:05 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:05Z [INFO] - [mgmt] - - - [2022-06-08 09:13:05.748824404 UTC] "GET /modules?api-version=2021-12-07 HTTP/1.1" 200 OK 1952 "-" "-" auth_id(-)
Jun 08 09:13:10 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:10Z [INFO] - [mgmt] - - - [2022-06-08 09:13:10.787328141 UTC] "GET /modules?api-version=2021-12-07 HTTP/1.1" 200 OK 1952 "-" "-" auth_id(-)
Jun 08 09:13:15 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:15Z [INFO] - [mgmt] - - - [2022-06-08 09:13:15.820943495 UTC] "GET /modules?api-version=2021-12-07 HTTP/1.1" 200 OK 1952 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (74 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /key?api-version=2020-09-01 {"content-length": "74", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 377 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (2599 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /decrypt?api-version=2020-09-01 {"content-length": "2599", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 2286 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [work] - - - [2022-06-08 09:13:20.484697762 UTC] "POST /modules/%24edgeHub/genid/637800092105391013/decrypt?api-version=2019-01-30 HTTP/1.1" 200 OK 2176 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (74 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /key?api-version=2020-09-01 {"content-length": "74", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 377 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (2574 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /encrypt?api-version=2020-09-01 {"content-length": "2574", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 2311 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [work] - - - [2022-06-08 09:13:20.564417345 UTC] "POST /modules/%24edgeHub/genid/637800092105391013/encrypt?api-version=2019-01-30 HTTP/1.1" 200 OK 2201 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (74 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /key?api-version=2020-09-01 {"content-length": "74", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 377 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (2599 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /decrypt?api-version=2020-09-01 {"content-length": "2599", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 2286 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [work] - - - [2022-06-08 09:13:20.640107437 UTC] "POST /modules/%24edgeHub/genid/637800092105391013/decrypt?api-version=2019-01-30 HTTP/1.1" 200 OK 2176 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (74 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /key?api-version=2020-09-01 {"content-length": "74", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 377 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (3614 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /encrypt?api-version=2020-09-01 {"content-length": "3614", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 3351 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [work] - - - [2022-06-08 09:13:20.717824087 UTC] "POST /modules/%24edgeHub/genid/637800092105391013/encrypt?api-version=2019-01-30 HTTP/1.1" 200 OK 3241 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (74 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /key?api-version=2020-09-01 {"content-length": "74", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 377 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (3639 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /decrypt?api-version=2020-09-01 {"content-length": "3639", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 3326 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [work] - - - [2022-06-08 09:13:20.806012386 UTC] "POST /modules/%24edgeHub/genid/637800092105391013/decrypt?api-version=2019-01-30 HTTP/1.1" 200 OK 3216 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [mgmt] - - - [2022-06-08 09:13:20.855975590 UTC] "GET /modules?api-version=2021-12-07 HTTP/1.1" 200 OK 1952 "-" "-" auth_id(-)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (74 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /key?api-version=2020-09-01 {"content-length": "74", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 377 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] parsed 2 headers
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body is content-length (3639 bytes)
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - <-- POST /decrypt?api-version=2020-09-01 {"content-length": "3639", "content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] incoming body completed
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [INFO] - --> 200 {"content-type": "application/json"}
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::io] flushed 3326 bytes
Jun 08 09:13:20 iot-gate-imx8 aziot-keyd[21666]: 2022-06-08T09:13:20Z [DBUG] - [hyper::proto::h1::conn] read eof
Jun 08 09:13:20 iot-gate-imx8 aziot-edged[21652]: 2022-06-08T09:13:20Z [INFO] - [work] - - - [2022-06-08 09:13:20.879883540 UTC] "POST /modules/%24edgeHub/genid/637800092105391013/decrypt?api-version=2019-01-30 HTTP/1.1" 200 OK 3216 "-" "-" auth_id(-)
As you can see, no message is received on the IoT Hub... I normally have a function app that will use the fallback route to receive messages sent.
Any way to investiguate it further on IoT Hub side ? Or some additional logs to investigate on python side ?
Thanks
@MrVhek have u raised this issue in the Edge repo?
To get more debug logs add:
import logging
logging.basicConfig(level=logging.DEBUG)
To this at the beginning of your application
@MrVhek have u raised this issue in the Edge repo?
Not yet, not sure if this can be related to your SDK or to Azure IoT Edge directly ?
To get more debug logs add:
import logging logging.basicConfig(level=logging.DEBUG)
At the beginning of your application.
Thanks, I've added it but don't see more python logs than those that I posted above.
I need to see if I can have more logs on Azure side.
i am facing same problem
@MrVhek can you try reverting to IoT Edge 1.1? There are known issues with using Edge 1.2 with the SDK.
Sorry for the late reply, I finally found the issue, it was related to the routing of send_message.
Be sure to not have an output after your module name (here GWApp or ToolsApp).
| gharchive/issue | 2022-06-08T13:34:29 | 2025-04-01T06:36:45.059688 | {
"authors": [
"MrVhek",
"anthonyvercolano",
"cartertinney",
"raowaqas72"
],
"repo": "Azure/azure-iot-sdk-python",
"url": "https://github.com/Azure/azure-iot-sdk-python/issues/1009",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1077223829 | chore: Updated language classifiers
Formally indicated support for Python 3.9 and 3.10
Formally indicated support for Python 3.9 and 3.10
| gharchive/pull-request | 2021-12-10T20:58:31 | 2025-04-01T06:36:45.061183 | {
"authors": [
"cartertinney"
],
"repo": "Azure/azure-iot-sdk-python",
"url": "https://github.com/Azure/azure-iot-sdk-python/pull/920",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1244313941 | Container Instances should support environment variable names containing dot '.'
System Info
OS: Windows 10
Shell: Ubuntu 20.04 on Windows
Azure CLI: 2.36.0
Expected Behavior
Executing the following Azure CLI command should create a container instance running Elasticsearch:
az container create --image elasticsearch:7.3.2 --name es -g test-rg --memory 2 --cpu 1 --ports 9200 --environment-variables "discovery.type=single-node"
Using environment variable names containing dot '.' should work.
Current Behavior
Executing the Azure CLI command exit with an error:
(InvalidContainerEnvironmentVariable) The environment variable name in container 'es' of container group 'es' is invalid. A valid environment variable name must start with alphabetic character or '_', followed by a string of alphanumeric characters or '_' (e.g. 'my_name', or 'MY_NAME', or 'MyName').
Code: InvalidContainerEnvironmentVariable
Message: The environment variable name in container 'es' of container group 'es' is invalid. A valid environment variable name must start with alphabetic character or '_', followed by a string of alphanumeric characters or '_' (e.g. 'my_name', or 'MY_NAME', or 'MyName').
@stephanwehr Thanks for reaching out to us and sharing this feedback. I was able to reproduce this issue from AzCLI as well as Powershell. Post isolating this issue, I found that this issue isn't related to the Container Instance but the underlying image OS type.
The bash manual mentions this clearly (here)
name
A word consisting solely of letters, numbers, and underscores, and beginning with a letter or underscore. Names are used as shell variable and function names. Also referred to as an identifier.
The elasticSearch image seems to be using Linux OSType. So the fix / workaround has to come from the OS side and not from the AzCLI / REST API Spec. Hope this helps.
@navba-MSFT you are correct, Bash can't but Elasticsearch can.
With the "env" command it is possible to set environment variables containing special characters.
Here is an example:
# env -i "cluster.initial_master_nodes=elelastico" "my:test:var=value" /bin/bash -c "env"
my:test:var=value
cluster.initial_master_nodes=elelastico
PWD=/root
SHLVL=0
_=/usr/bin/env
In Docker compose the environment can be set like this:
version: '2.2'
services:
elasticsearch:
image: elasticsearch:7.3.2
environment:
- "ES_JAVA_OPTS=-Xms512m -Xmx512m"
- discovery.type=single-node
ports:
- 9200:9200
I've dug a little bit deeper on this issues and it seems Linux executes bash and other programs with execve().
With environ() an initial environment can be set.
How Linux execute programs can be seen using "strace":
$ strace env -i "my:test:var=some text" /bin/bash -c ":" 2>&1 | grep "execve"
execve("/usr/bin/env", ["env", "-i", "my:test:var=some text", "/bin/bash", "-c", ":"], 0x7ffd31dcb018 /* 21 vars */) = 0
execve("/bin/bash", ["/bin/bash", "-c", ":"], 0x5577895c20f0 /* 1 var */) = 0
As I understand, on Linux container environments like Docker for example set the initial environment before executing the 1st program inside the container. There is a process environment and bash environment which both can be retrieved using commands like "env" or "printenv".
I assume referring to this error message that the API performs a syntax check on variable names which should not be there:
{"error":{"code":"InvalidContainerEnvironmentVariable","message":"The environment variable name in container 'es' of container group 'es' is invalid. A valid environment variable name must start with alphabetic character or '_', followed by a string of alphanumeric characters or '_' (e.g. 'my_name', or 'MY_NAME', or 'MyName')."}}
@stephanwehr I just wanted to confirm if you are able to get this working with elasticsearch image with Linux OStype ? Awaiting your reply.
@navba-MSFT I'm not able to get Elasticsearch to work with Linux OStype because environment variables containing dot '.' are not supported.
@stephanwehr AFAIK, this ask is something which violates the IEEE standard for environment variables.
Environment variable names used by the utilities in the Shell and Utilities volume of IEEE Std 1003.1-2001 consist solely of uppercase letters, digits, and the '_' (underscore)
Refer this.
@navba-MSFT as stated this applies to the Shell and Utilities and not necessarily to other programs like Elasticsearch.
There is no hard requirement under Linux or inside a Docker container to use a shell to execute programs.
Linux itself allows variable names of type char:
#include <stdlib.h>
int setenv(const char *name, const char *value, int overwrite);
int unsetenv(const char *name);
Reference
@stephanwehr Thanks for clarifying. I have added the Service Team to look into this feature request ask.
I've tried hosting ElasticSearch on ACI and got this issue.
Are there any updates?
Same for me tried to use ElasticSearch with ACI and I get the same error with dotted env variable names. Why is this closed? There is no fix or workaround
| gharchive/issue | 2022-05-22T17:10:44 | 2025-04-01T06:36:45.151597 | {
"authors": [
"iwate",
"justinpenguin45",
"navba-MSFT",
"stephanwehr"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/issues/19173",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
496459502 | webapp:New-AzResourceGroupDeployment for web site with slot fails if template contains "cloningInfo"
Description
I'm trying to deploy a web app with two slots (production, deploytarget). Everything works fine the first time, but subsequent runs with the web app already present fail with an error indicating the site already exists. I was under the impression that ARM templates should always be able to run even if the resources already exist? If not is there a quick way to tell ARM that it should not touch the slot if it exists?
The problem goes away if I remove the "cloningInfo" options from the template. It appears there is a problem with cloning onto an existing app service.
"cloningInfo": {
"sourceWebAppId": "[resourceId('Microsoft.Web/sites', variables('webAppName'))]",
"overwrite": true
}
Script/Steps for Reproduction
New-AzResourceGroupDeployment # with provided template
This is the template that is used:
{
"$schema": "https://schema.management.azure.com/schemas/2015-01-01/deploymentTemplate.json#",
"contentVersion": "1.0.0.0",
"parameters": {
"environment": {
"type": "string",
"allowedValues": [
"devtest",
"test",
"stag",
"prod"
]
},
"skuName": {
"type": "string",
"defaultValue": "S1"
},
"skuCapacity": {
"type": "int",
"defaultValue": 1
},
"nodeVersion": {
"type": "string",
"defaultValue": "10.6.0"
},
"location": {
"type": "string",
"defaultValue": "[resourceGroup().location]"
},
"deployTargetSlotName": {
"type": "string",
"defaultValue": "deploytarget"
}
},
"variables": {
"appServiceName": "[concat('TestAppServiceName-ASP-', parameters('environment'))]",
"webAppName": "[concat('TestAppServiceName-APP-', parameters('environment'))]"
},
"resources": [
{
"name": "[variables('appServiceName')]",
"type": "Microsoft.Web/serverfarms",
"kind": "app",
"sku": {
"name": "[parameters('skuName')]",
"capacity": "[parameters('skuCapacity')]"
},
"apiVersion": "2016-09-01",
"location": "[parameters('location')]",
"scale": null,
"properties": {
"name": "[variables('appServiceName')]"
}
},
{
"name": "[variables('webAppName')]",
"type": "Microsoft.Web/sites",
"apiVersion": "2018-02-01",
"location": "[parameters('location')]",
"tags": {
"displayName": "[variables('webAppName')]"
},
"properties": {
"name": "[variables('webAppName')]",
"serverFarmId": "[resourceId('Microsoft.Web/serverfarms', variables('appServiceName'))]",
"clientAffinityEnabled": false,
"httpsOnly": true,
"siteConfig": {
"alwaysOn": true,
"phpVersion": ""
}
},
"dependsOn": [
"[variables('appServiceName')]"
],
"resources": [
{
"name": "[parameters('deployTargetSlotName')]",
"type": "slots",
"apiVersion": "2018-02-01",
"dependsOn": [
"[variables('webAppName')]"
],
"location": "[parameters('location')]",
"properties": {
"enabled": false,
"serverFarmId": "[resourceId('Microsoft.Web/serverfarms', variables('appServiceName'))]",
"clientAffinityEnabled": false,
"httpsOnly": true,
"cloningInfo": {
"sourceWebAppId": "[resourceId('Microsoft.Web/sites', variables('webAppName'))]",
"overwrite": true
}
}
}
]
}
],
"outputs": {
"webappName": {
"type": "string",
"value": "[variables('webAppName')]"
}
}
}
Module Version
ModuleType Version Name ExportedCommands
---------- ------- ---- ----------------
Script 0.6.1 Az.Profile {Add-AzEnvironment, Clear-AzContext, Clear-AzDefault, Connect-AzAccount...}
Script 0.6.1 Az.Storage {Add-AzRmStorageContainerLegalHold, Add-AzStorageAccountNetworkRule, Disable-AzStorageDeleteRetentionPolicy, Enable-AzStorageDeleteRetentionPolicy...}
Manifest 3.1.0.0 Microsoft.PowerShell.Management {Add-Computer, Add-Content, Checkpoint-Computer, Clear-Content...}
Manifest 3.0.0.0 Microsoft.PowerShell.Security {ConvertFrom-SecureString, ConvertTo-SecureString, Get-Acl, Get-AuthenticodeSignature...}
Manifest 3.1.0.0 Microsoft.PowerShell.Utility {Add-Member, Add-Type, Clear-Variable, Compare-Object...}
Manifest 3.0.0.0 Microsoft.WSMan.Management {Connect-WSMan, Disable-WSManCredSSP, Disconnect-WSMan, Enable-WSManCredSSP...}
Script 1.2 PSReadline {Get-PSReadlineKeyHandler, Get-PSReadlineOption, Remove-PSReadlineKeyHandler, Set-PSReadlineKeyHandler...}
Environment Data
Name Value
---- -----
PSVersion 5.1.17134.407
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.17134.407
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
Debug Output
13:23:59 - Resource Microsoft.Web/sites/slots 'TestAppServiceName-ASP-devtest/deploytarget' failed with message '{
"status": "Failed",
"error": {
"code": "ResourceDeploymentFailure",
"message": "The resource operation completed with terminal provisioning state 'Failed'.",
"details": [
{
"message": "Conflict\r\nDetail: Site with name TestAppServiceName-ASP-devtest(deploytarget) already exists.\r\nExtendedCode: 04008\r\n at Microsoft.Web.Hosting.Administration.Csm.Common.Clients.GeoClientBase.GetResponse[T](HttpWebRequest request, Boolean reThrowOnHostUnavailability, HttpStatusCode[]
supressLogForHttpCodes)\r\n at Microsoft.Web.Hosting.Administration.GeoScale.Sql.WebSiteCloneManager.StartSiteRestore(SiteClone clone, Operation cloneOperation)\r\n"
}
]
}
}'
Is there any update in regards to this issue?
@panchagnula so its been over 2 years now that this bug was reported. You reassigned this over a year ago, but nothing happened.
@zikalino any update?
@maggiepint you maybe?
Can anyone from microsoft please give a status update?
@rudfoss did you ever find a workaround?
this issue seems already fixed
Is this issue ever going to be fixed? It is impossible to redeploy a template when using cloningInfo unless you know ahead of time if the resource already exists and set the overwrite property accordingly. This goes against the whole premise of ARM.
| gharchive/issue | 2018-12-04T12:38:55 | 2025-04-01T06:36:45.159719 | {
"authors": [
"PaulVrugt",
"m1dst",
"rudfoss",
"simonethreechillies"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/issues/7276",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1222819881 | Fixing swagger correctness items RSV
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Changelog
Add a changelog entry for this PR by answering the following questions:
What's the purpose of the update?
[ ] new service onboarding
[ ] new API version
[ ] update existing version for new feature
[ ] update existing version to fix swagger quality issue in s360
[ ] Other, please clarify
When are you targeting to deploy the new service/feature to public regions? Please provide the date or, if the date is not yet available, the month.
When do you expect to publish the swagger? Please provide date or, the the date is not yet available, the month.
If updating an existing version, please select the specific language SDKs and CLIs that must be refreshed after the swagger is published.
[ ] SDK of .NET (need service team to ensure code readiness)
[ ] SDK of Python
[ ] SDK of Java
[ ] SDK of Js
[ ] SDK of Go
[ ] PowerShell
[ ] CLI
[ ] Terraform
[ ] No refresh required for updates in this PR
Contribution checklist:
[ ] I commit to follow the Breaking Change Policy of "no breaking changes"
[ ] I have reviewed the documentation for the workflow.
[ ] Validation tools were run on swagger spec(s) and errors have all been fixed in this PR. How to fix?
If any further question about AME onboarding or validation tools, please view the FAQ.
ARM API Review Checklist
Applicability: :warning:
If your changes encompass only the following scenarios, you should SKIP this section, as these scenarios do not require ARM review.
Change to data plane APIs
Adding new properties
All removals
Otherwise your PR may be subject to ARM review requirements. Complete the following:
[ ] Check this box if any of the following apply to the PR so that label "WaitForARMFeedback" will be added automatically to begin ARM API Review. Failure to comply may result in delays to the manifest.
Adding a new service
Adding new API(s)
Adding a new API version
-[ ] To review changes efficiently, ensure you are using OpenAPIHub to initialize the PR for adding a new version. More details, refer to the wiki.
[ ] Ensure you've reviewed following guidelines including ARM resource provider contract and REST guidelines. Estimated time (4 hours). This is required before you can request review from ARM API Review board.
[ ] If you are blocked on ARM review and want to get the PR merged with urgency, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Breaking Change Review Checklist
If any of the following scenarios apply to the PR, request approval from the Breaking Change Review Board as defined in the Breaking Change Policy.
[ ] Removing API(s) in a stable version
[ ] Removing properties in a stable version
[ ] Removing API version(s) in a stable version
[ ] Updating API in a stable or public preview version with Breaking Change Validation errors
[ ] Updating API(s) in public preview over 1 year (refer to Retirement of Previews)
Action: to initiate an evaluation of the breaking change, create a new intake using the template for breaking changes. Addition details on the process and office hours are on the Breaking change Wiki.
Please follow the link to find more details on PR review process.
Hi @pratimaupadhyay02, please add the new properties to new version and fix the linter errors.
| gharchive/pull-request | 2022-05-02T12:20:50 | 2025-04-01T06:36:45.175375 | {
"authors": [
"pratimaupadhyay02",
"xiaoxuqi-ms"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/18875",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1233138482 | Copy QuestionAnswering definitions to common.json and refer them
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Changelog
Add a changelog entry for this PR by answering the following questions:
What's the purpose of the update?
[ ] new service onboarding
[ ] new API version
[ ] update existing version for new feature
[ ] update existing version to fix swagger quality issue in s360
[ ] Other, please clarify
When are you targeting to deploy the new service/feature to public regions? Please provide the date or, if the date is not yet available, the month.
When do you expect to publish the swagger? Please provide date or, the the date is not yet available, the month.
If updating an existing version, please select the specific language SDKs and CLIs that must be refreshed after the swagger is published.
[ ] SDK of .NET (need service team to ensure code readiness)
[ ] SDK of Python
[ ] SDK of Java
[ ] SDK of Js
[ ] SDK of Go
[ ] PowerShell
[ ] CLI
[ ] Terraform
[ ] No refresh required for updates in this PR
Contribution checklist:
[ ] I commit to follow the Breaking Change Policy of "no breaking changes"
[ ] I have reviewed the documentation for the workflow.
[ ] Validation tools were run on swagger spec(s) and errors have all been fixed in this PR. How to fix?
If any further question about AME onboarding or validation tools, please view the FAQ.
ARM API Review Checklist
Applicability: :warning:
If your changes encompass only the following scenarios, you should SKIP this section, as these scenarios do not require ARM review.
Change to data plane APIs
Adding new properties
All removals
Otherwise your PR may be subject to ARM review requirements. Complete the following:
[ ] Check this box if any of the following appy to the PR so that the label "ARMReview" and "WaitForARMFeedback" will be added by bot to kick off ARM API Review. Missing to check this box in the following scenario may result in delays to the ARM manifest review and deployment.
Adding a new service
Adding new API(s)
Adding a new API version
-[ ] To review changes efficiently, ensure you are using OpenAPIHub to initialize the PR for adding a new version. More details, refer to the wiki.
[ ] Ensure you've reviewed following guidelines including ARM resource provider contract and REST guidelines. Estimated time (4 hours). This is required before you can request review from ARM API Review board.
[ ] If you are blocked on ARM review and want to get the PR merged with urgency, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Breaking Change Review Checklist
If any of the following scenarios apply to the PR, request approval from the Breaking Change Review Board as defined in the Breaking Change Policy.
[ ] Removing API(s) in a stable version
[ ] Removing properties in a stable version
[ ] Removing API version(s) in a stable version
[ ] Updating API in a stable or public preview version with Breaking Change Validation errors
[ ] Updating API(s) in public preview over 1 year (refer to Retirement of Previews)
Action: to initiate an evaluation of the breaking change, create a new intake using the template for breaking changes. Addition details on the process and office hours are on the Breaking change Wiki.
Please follow the link to find more details on PR review process.
@lmazuel, the only breaking change is our old friend the now-required error property. Could you merge regardless?
Hopefully once this GAs to main this specific issue will be a thing of the past.
@lmazuel, the only breaking change is our old friend the now-required error property. Could you merge regardless?
Hopefully once this GAs to main this specific issue will be a thing of the past.
Hi @lmazuel , can you help us merge it? Thanks!
| gharchive/pull-request | 2022-05-11T20:17:51 | 2025-04-01T06:36:45.193617 | {
"authors": [
"ChongTang",
"heaths"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/19030",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2384396672 | [Batch] Switch contentLength from int64 to string
This is a breaking change, but accurately reflects what the Batch service has always returned for this property.
Data Plane API Specification Update Pull Request
[!TIP]
Overwhelmed by all this guidance? See the Getting help section at the bottom of this PR description.
PR review workflow diagram
Please understand this diagram before proceeding. It explains how to get your PR approved & merged.
API Info: The Basics
Most of the information about your service should be captured in the issue that serves as your API Spec engagement record.
Link to API Spec engagement record issue:
Is this review for (select one):
[ ] a private preview
[x] a public preview
[ ] GA release
Change Scope
This section will help us focus on the specific parts of your API that are new or have been modified. Please share a link to the design document for the new APIs, a link to the previous API Spec document (if applicable), and the root paths that have been updated.
Design Document:
Previous API Spec Doc:
Updated paths:
Viewing API changes
For convenient view of the API changes made by this PR, refer to the URLs provided in the table
in the Generated ApiView comment added to this PR. You can use ApiView to show API versions diff.
Suppressing failures
If one or multiple validation error/warning suppression(s) is detected in your PR, please follow the
Swagger-Suppression-Process
to get approval.
❔Got questions? Need additional info?? We are here to help!
Contact us!
The Azure API Review Board is dedicated to helping you create amazing APIs. You can read about our mission and learn more about our process on our wiki.
💬 Teams Channel
💌 email
Click here for links to tools, specs, guidelines & other good stuff
Tooling
Open API validation tools were run on this PR. Go here to see how to fix errors
Spectral Linting
Guidelines & Specifications
Azure REST API Guidelines
OpenAPI Style Guidelines
Azure Breaking Change Policy
Helpful Links
Schedule a data plane REST API spec review
Getting help
First, please carefully read through this PR description, from top to bottom.
If you don't have permissions to remove or add labels to the PR, request write access per aka.ms/azsdk/access#request-access-to-rest-api-or-sdk-repositories
To understand what you must do next to merge this PR, see the Next Steps to Merge comment. It will appear within few minutes of submitting this PR and will continue to be up-to-date with current PR state.
For guidance on fixing this PR CI check failures, see the hyperlinks provided in given failure
and https://aka.ms/ci-fix.
If the PR CI checks appear to be stuck in queued state, please add a comment with contents /azp run.
This should result in a new comment denoting a PR validation pipeline has started and the checks should be updated after few minutes.
If the help provided by the previous points is not enough, post to https://aka.ms/azsdk/support/specreview-channel and link to this PR.
Note that this TypeSpec is currently in a preview state.
/azp run
@raych1 is there anything blocking this from being merged? I'm not sure what the VersionReviewRequired tag is, but if it's regarding making changes to an existing API version, I think we should be fine. The TypeSpec in this PR is in a public preview state, not GA.
/azp run
/azp run
| gharchive/pull-request | 2024-07-01T17:37:30 | 2025-04-01T06:36:45.209048 | {
"authors": [
"dpwatrous",
"raych1"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/29656",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
424334140 | typo: consumption/resource-manager/Microsoft.Consumption
chages -> charges
Cylce -> Cycle
Can one of the admins verify this patch?
Automation for azure-sdk-for-js
Automation for azure-sdk-for-python
Automation for azure-sdk-for-ruby
Automation for azure-sdk-for-node
Automation for azure-sdk-for-go
Automation for azure-sdk-for-java
@sandeepkatumalla could you take a look since you worked on this version last?
I think the build failure is actually a message issue https://github.com/Azure/avocado/issues/11
| gharchive/pull-request | 2019-03-22T18:01:01 | 2025-04-01T06:36:45.213232 | {
"authors": [
"AutorestCI",
"azuresdkci",
"dsgouda",
"nschonni"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/5458",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
508205631 | Updated ContentMD5 parameter in DataLake swagger
Latest improvements:
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Contribution checklist:
[ ] I have reviewed the documentation for the workflow.
[ ] Validation tools were run on swagger spec(s) and have all been fixed in this PR.
[ ] The OpenAPI Hub was used for checking validation status and next steps.
ARM API Review Checklist
[ ] Service team MUST add the "WaitForARMFeedback" label if the management plane API changes fall into one of the below categories.
adding/removing APIs.
adding/removing properties.
adding/removing API-version.
adding a new service in Azure.
Failure to comply may result in delays for manifest application. Note this does not apply to data plane APIs.
[ ] If you are blocked on ARM review and want to get the PR merged urgently, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Please follow the link to find more details on API review process.
Can one of the admins verify this patch?
Automation for azure-sdk-for-python
Automation for azure-sdk-for-go
| gharchive/pull-request | 2019-10-17T03:09:02 | 2025-04-01T06:36:45.219036 | {
"authors": [
"AutorestCI",
"azuresdkci",
"seanmcc-msft"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/7534",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
591968682 | Managed Instance failover swagger
Latest improvements:
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Contribution checklist:
[x] I have reviewed the documentation for the workflow.
[x] Validation tools were run on swagger spec(s) and have all been fixed in this PR.
[ ] The OpenAPI Hub was used for checking validation status and next steps.
ARM API Review Checklist
[x] Service team MUST add the "WaitForARMFeedback" label if the management plane API changes fall into one of the below categories.
adding/removing APIs.
adding/removing properties.
adding/removing API-version.
adding a new service in Azure.
Failure to comply may result in delays for manifest application. Note this does not apply to data plane APIs.
[ ] If you are blocked on ARM review and want to get the PR merged urgently, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Please follow the link to find more details on API review process.
/azp run automation - sdk
Can one of the admins verify this patch?
/azp run automation - sdk
/azp run automation - sdk
/azp run automation - sdk
/azp run automation - sdk
/azp run
@MilanBrkicFON Any update on this PR?
@NullMDR Last week changes landed in production, but I didn't have time to update this pr.
Now is ready for review!
/azp run
/azp run automation - sdk
Hi @NullMDR are you ok with the changes? Can we merge this PR?
/azp run
@MilanBrkicFON You need to wait for ARM review. I've emailed them.
/azp run automation - sdk
/azp run
@pilor @allenjzhang Can I have this PR reviewed please?
@pilor @allenjzhang Can I have this PR reviewed please?
API reviews are a weekly rotation described here. Please reach out to this week or next week's reviewer
@MilanBrkicFON, pretty straightforward PR, though I have noticed consistent pattern issues with SQL swagger files. Please take a look at these easy fixes. Will loop back to work with you to cleanup if these are manually created. If these are generated, please incorporate feedback and make appropriate changes to generate higher quality swagger in the future.
BTW, this is SDK team review. As per policy, ARM folks need to sign off for new APIs. Giving the low complexity of this PR, it should be a quick sign off. Please engage ARM folks per @pilor's instruction.
/azp run automation - sdk
/azp run automation - sdk
@allenjzhang Thanks for the review. I think I've addressed all of your comments. Can you please check again?
I will also update internal docs on ARM development to warn engineers what we are doing wrong and to avoid same mistakes in the future, but I think the right solution here would be working auto-generation script.
/azp run automation - sdk
/azp run
/azp run automation - sdk
/azp run automation - sdk
@pilor @allenjzhang @NullMDR Can someone merge this PR please? I am not authorized to do it.
| gharchive/pull-request | 2020-04-01T14:39:52 | 2025-04-01T06:36:45.232302 | {
"authors": [
"AutorestCI",
"MilanBrkicFON",
"NullMDR",
"allenjzhang",
"azuresdkci",
"pilor"
],
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/8911",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
967553270 | Investigate telemetry possibly not getting in?
From @kyle-patterson:
"[...] I can see that for storage up through -beta.9, but there's no new data coming in from libraries since then. And for keyvault, I can only see two instances of "azsdk-cpp-keyvault/7.2 {os info...}" , where I would have expected keyvault-keys (and more data coming through...) [...]
It almost seems like there was a change in core in the june or july release, but I haven't found anything yet that seems related...
I did confirm that keyvault is passing the expected value when creating the ttp pipeline: https://github.com/Azure/azure-sdk-for-cpp/blob/83295c69edc5d2a6594c09fcd06eae7073753171/sdk/keyvault/azure-security-keyvault-keys/src/key_client.cpp#L28
it's not hyper-critical, but I'd like to get it resolved in the next release so that we can start getting data again..."
I wrote a simple app that invokes key creation API.
In order to avoid setting up HTTPS decryption on my local machine, I modified the SDK to send requests to a different server on the internet instead of Azure, and read headers there. I compiled with both libcurl and WinHTTP transport adapters. Both have azsdk-cpp-keyvault-keys/7.2 (Windows 10 Enterprise 6.3 19043 19041.1.amd64fre.vb_release.191206-1406) as User-Agent header. (Note that the version does not match the package version, I opened a bug for that - https://github.com/Azure/azure-sdk-for-cpp/issues/2765). But other than that, it look like everything works as it supposed to.
@kyle-patterson can see the telemetry now, I won't post much detail here (not that there's something secret-worthy or sensational was found). Kyle, please reactivate or talk to me if you see problems.
| gharchive/issue | 2021-08-11T22:11:29 | 2025-04-01T06:36:45.236637 | {
"authors": [
"antkmsft"
],
"repo": "Azure/azure-sdk-for-cpp",
"url": "https://github.com/Azure/azure-sdk-for-cpp/issues/2738",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1037098577 | Feature request: Core SDK should support HTTP request with request body and non-buffered(stream) response
Currently we have three constructor overloads:
with request body, buffered response
without request body, non-buffered response
without request body, buffered response
Storage SDK needs HTTP request with request body and non-buffered(stream) response to implement Query Blob Content
reopen since it was reverted in https://github.com/Azure/azure-sdk-for-cpp/pull/3033
@vhvb1989 Can you share a timeline for fixing this issue?
@vhvb1989 Can you share a timeline for fixing this issue?
I did in the past with: https://github.com/Azure/azure-sdk-for-cpp/pull/3002
However, I reverted it as it was not yet required from storage. So, feel free to re-apply the changes from that PR together with the Storage feature that it requires it.
OK, that works for me.
| gharchive/issue | 2021-10-27T08:00:50 | 2025-04-01T06:36:45.240824 | {
"authors": [
"Jinming-Hu",
"vhvb1989"
],
"repo": "Azure/azure-sdk-for-cpp",
"url": "https://github.com/Azure/azure-sdk-for-cpp/issues/3000",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
838244939 | mint storage beta 9
Pull Request Checklist
Please leverage this checklist as a reminder to address commonly occurring feedback when submitting a pull request to make sure your PR can be reviewed quickly:
See the detailed list in the contributing guide.
[x] C++ Guidelines
[x] Doxygen docs
[x] Unit tests
[x] No unwanted commits/changes
[x] Descriptive title/description
[x] PR is single purpose
[x] Related issue listed
[x] Comments in source
[x] No typos
[x] Update changelog
[x] Not work-in-progress
[x] External references or docs updated
[x] Self review of PR done
[x] Any breaking changes?
/azp run cpp - storage
/azp run cpp - storage
/azp run cpp - storage
| gharchive/pull-request | 2021-03-23T01:35:20 | 2025-04-01T06:36:45.245661 | {
"authors": [
"Jinming-Hu"
],
"repo": "Azure/azure-sdk-for-cpp",
"url": "https://github.com/Azure/azure-sdk-for-cpp/pull/1957",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1260195432 | Removed version>= fields for openssl in vcpkg.json files
Remove version>= fields from vcpkg.json files.
Fixes #3703
These fields are broken because:
These fields are illegal in the vcpkg schema because version>= only works with semver version numbers and 1.1.1n is not a legal semver version number.
We rely on the vcpkg baseline to determine which version of openssl we use, currently 3.0.2.
Pull Request Checklist
Please leverage this checklist as a reminder to address commonly occurring feedback when submitting a pull request to make sure your PR can be reviewed quickly:
See the detailed list in the contributing guide.
[X] C++ Guidelines
[X] Doxygen docs
[X] Unit tests
[X] No unwanted commits/changes
[X] Descriptive title/description
[X] PR is single purpose
[X] Related issue listed
[X] Comments in source
[X] No typos
[X] Update changelog
[X] Not work-in-progress
[X] External references or docs updated
[X] Self review of PR done
[X] Any breaking changes?
Would it be best to have version >= 3.0.2 to ensure its at least of that version?
Would it be best to have version >= 3.0.2 to ensure its at least of that version?
A great question, not 100% sure to be honest. We define the version we take from our vcpkg baseline (eng\vcpkg-commit.txt) and we'll pick the version of whatever vcpkg has at that point. As I understand it, the version overrides are there if we specifically require functionality which might be different from the baseline (typically older than the baseline).
If we just want to use whatever the baseline has, we should leave it blank.
| gharchive/pull-request | 2022-06-03T17:57:49 | 2025-04-01T06:36:45.252779 | {
"authors": [
"LarryOsterman",
"RickWinter"
],
"repo": "Azure/azure-sdk-for-cpp",
"url": "https://github.com/Azure/azure-sdk-for-cpp/pull/3705",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
642043652 | package should be named "avs", not "vmware"
The package should be named "avs", not "vmware". #10454 was merged before I had a chance to review. Please rename the package. See also https://github.com/Azure/sdk-release-request/issues/496.
Current location:
https://github.com/Azure/azure-sdk-for-go/tree/master/services/preview/vmware/mgmt/2019-08-09-preview/vmware
Should be:
https://github.com/Azure/azure-sdk-for-go/tree/master/services/preview/avs/mgmt/2019-08-09-preview/avs
Hi @ctaggart sorry for that... I will change it, release it and remove the wrong-named package in the next major version release (at the end of this month)
Should be resolved in v44.0.0
thank you @ArcturusZhang ! ❤️
| gharchive/issue | 2020-06-19T15:15:06 | 2025-04-01T06:36:45.255975 | {
"authors": [
"ArcturusZhang",
"ctaggart"
],
"repo": "Azure/azure-sdk-for-go",
"url": "https://github.com/Azure/azure-sdk-for-go/issues/10683",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
940025381 | [Tables] Shared Access Signature Capabilities
Adding SAS capabilities to the Tables SDK
[ ] Add Shared Access Signature credential
[ ] SharedAccessSignaturePolicy for signing requests properly
[ ] generate_account_sas method
[ ] generate_table_sas method
[ ] TableSasPermission object
[ ] AccountSasPermissions object
[ ] ResourceTypes object
[ ] AccountSasPermissions object
Python implementation: https://github.com/Azure/azure-sdk-for-python/pull/15946
| gharchive/issue | 2021-07-08T16:22:01 | 2025-04-01T06:36:45.259482 | {
"authors": [
"seankane-msft"
],
"repo": "Azure/azure-sdk-for-go",
"url": "https://github.com/Azure/azure-sdk-for-go/issues/15004",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
865267964 | Sync eng/common directory with azure-sdk-tools for PR 1562
Sync eng/common directory with azure-sdk-tools for PR https://github.com/Azure/azure-sdk-tools/pull/1562 See eng/common workflow
/check-enforcer reset
| gharchive/pull-request | 2021-04-22T18:03:07 | 2025-04-01T06:36:45.260958 | {
"authors": [
"azure-sdk",
"weshaggard"
],
"repo": "Azure/azure-sdk-for-go",
"url": "https://github.com/Azure/azure-sdk-for-go/pull/14581",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
972002126 | Fixed bugs in Resource Group demo code
[ ] The purpose of this PR is explained in this or a referenced issue.
[ ] The PR does not update generated files.
These files are managed by the codegen framework at Azure/autorest.go.
[ ] Tests are included and/or updated for code changes.
[ ] Updates to CHANGELOG.md are included.
[ ] MIT license headers are included in each file.
@RickWinter I believe I fixed the indentation.
| gharchive/pull-request | 2021-08-16T18:38:46 | 2025-04-01T06:36:45.263912 | {
"authors": [
"TomArcherMsft"
],
"repo": "Azure/azure-sdk-for-go",
"url": "https://github.com/Azure/azure-sdk-for-go/pull/15301",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
192905819 | route table / route / subnet association
Is there a path for managing route table / routes and their subnet association in beta3 ? The SDK points for route table creation, route management and subnet association in beta4 are great and are working perfectly. However, since I am currently required to use released or final code as dependencies, is there any way to manage these same things via beta3 or interaction with an ARM template (don't believe ARM templates allow for update only deploy)
not easily - there may be a workaround via the use of .inner() of the various involved objects, but that approach will get very hairy very quickly....
beta4 is coming out in about 1-2 weeks though, so better just wait for that.
Ahh beta4 is the answer I was hoping for and needed to know a timeline but was not sure where to ask. Thank you very much for the prompt respond, and I believe an ARM template can be used temporarily in the mean time.
| gharchive/issue | 2016-12-01T17:07:21 | 2025-04-01T06:36:45.265729 | {
"authors": [
"martinsawicki",
"paulojohnj"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/issues/1299",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
708523493 | Decouple PagedIterable and PagedFlux to Prevent Background Page Requests
Fixes #15575
This PR decouples some ContinuablePagedIterable implementations from ContinuablePagedFlux when the ContinuablePagedFlux implementation is an instanceof ContinuablePagedFluxCore. This change allows for finer control of page enumeration by removing our dependency on Reactor's toIterable and toStream implementations on a Flux.
Previously, when our implementation used Reactor's toIterable and toStream to implement ContinuablePagedIterable's functionality we would see additional pages getting requested. Calling into these methods we passed a batchSize of 1 which indicated to the backing Flux to only request one element, paged, from upstream at time. But this also served a dual purpose of internal tracking for the backing enumerable to determine when it needed to make additional requests to upstream, so on every next iteration it would hit its internal limit of 1 and make another page request. So, calling ContinuablePagedIterable.streamByPage().findFirst() or ContinuablePagedIterable.iterableByPage().iterator().next() would result in two page requests. Additionally, due to the reactive, event loop, driven nature of the backing enumerable these additional page requests could happen after the execution of the mentioned calling patterns completed.
Now, during construction of ContinuablePagedIterable we will check the instanceof the backing ContinuablePagedFlux and if it is ContinuablePagedFluxCore the PageRetriever and batchSize configuration will be taken from the object and ContinuablePagedIterable will handle page enumeration with finer grain controls. To maintain current functionality that Flux.toStream() has a page will be eagerly requested when a Iterator or Stream is created from the ContinuablePagedIterable. Internally, these will be backed by one of two Iterable implementations, ContinuablePagedByItemIterable or ContinuablePagedByPageIterable. The implementations will only make additional page requests when needed, for page item iterators it will be when the most recently retrieved page has no additional elements and for page iterators it will be once the next page is requested. The page requests will be blocking which is functionally equivalent to the previous experience, where both implementations would throw if called from within a non-blocking reactive thread.
In the future we may investigate completely decoupling ContinuablePagedIterable and ContinuablePagedFlux by having both be constructed with PagedRetriever and handle their own enumeration of pages.
/azp run java - appconfiguration - tests
| gharchive/pull-request | 2020-09-24T22:57:54 | 2025-04-01T06:36:45.271272 | {
"authors": [
"alzimmermsft"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/15646",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
736688332 | Added support for encryption algorithms for symmetric keys
Fixes #14805.
Tests are pending.
A few changes required yet, but I'm signing off to unblock. Also, LocalKeyCryptographyClient->LocalCryptographyClient if you haven't GA'd it yet.
We have already GA'd a public LocalCryptographyClient. In the case of LocalKeyCryptographyClient, it's just an abstract class that we extend from for our different internal clients for EC, RSA and AES key operations.
| gharchive/pull-request | 2020-11-05T07:33:39 | 2025-04-01T06:36:45.273127 | {
"authors": [
"vcolin7"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/17209",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1056043391 | Make the dependency versions the same with SDK Bom managed
As title.
API changes have been detected in com.azure:azure-core. You can review API changes here
API changes
+ public enum ErrorOptions {
+ THROW,
+ NO_THROW;
+ }
+ public RequestOptions setErrorOptions(EnumSet<ErrorOptions> errorOptions)
We decided not to use Azure SDK Bom first, so I close this PR now.
| gharchive/pull-request | 2021-11-17T11:57:04 | 2025-04-01T06:36:45.275414 | {
"authors": [
"azure-sdk",
"moarychan"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/25470",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1698236882 | STG88 Features
Description
This PR introduces service version support for 2023-01-03, and the following STG88 features:
High Throughput Append Blob for AppendBlobClient
File Share List Handles Access Rights
Add Owner, Group, and Permissions to PathProperties for DataLake
API change check
API changes are not detected in this pull request.
/azp run java - storage - tests
/check-enforcer override
/azp run java - storage - tests
/check-enforcer override
| gharchive/pull-request | 2023-05-05T21:56:23 | 2025-04-01T06:36:45.278238 | {
"authors": [
"alzimmermsft",
"azure-sdk",
"ibrahimrabab"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/34820",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
440885234 | [AutoPR graphrbac/data-plane] graph: fix invalid types of accountEnabled
Created to sync https://github.com/Azure/azure-rest-api-specs/pull/5873
This PR has been merged into https://github.com/Azure/azure-sdk-for-java/pull/3099
| gharchive/pull-request | 2019-05-06T21:01:06 | 2025-04-01T06:36:45.279960 | {
"authors": [
"AutorestCI"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/3554",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1984920412 | fix429EscapeE2ETimeout
Changes included:
Fixed Issue https://github.com/Azure/azure-sdk-for-java/issues/37419
Issue
429 escaped e2e timeout gating.
Root cause
Currently, the E2ETimeout for point operations is not sitting on top of ClientRetryPolicy, which means for each retry from ClientRetryPolicy will reset the e2e timeout timer.
Fix
For point operations, E2E timeout should be gated for all retries by client retry policy
Fixed Issue https://github.com/Azure/azure-sdk-for-java/issues/37589
Issue
E2E timeout is not being applied for readMany (Query cases)
Fix
Apply E2E timeout for readMany query as well
API change check
API changes are not detected in this pull request.
/azp run java - cosmos - tests
| gharchive/pull-request | 2023-11-09T06:25:09 | 2025-04-01T06:36:45.283870 | {
"authors": [
"azure-sdk",
"xinlian12"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/37591",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2593544038 | [Automation] Generate Fluent Lite from Swagger mariadb#package-2020-01-01
[Automation] Generate Fluent Lite from Swagger mariadb#package-2020-01-01
/azp run java - mariadb - ci
API change check
APIView has identified API level changes in this PR and created following API reviews.
com.azure.resourcemanager:azure-resourcemanager-mariadb
| gharchive/pull-request | 2024-10-17T03:58:02 | 2025-04-01T06:36:45.286043 | {
"authors": [
"XiaofeiCao",
"azure-sdk"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/42386",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
537304496 | Release changes for 3.4.1
Changefeed bug fixes (includes all the bug fixes released in 3.3.3)
Disaster Recovery related bug fixes
Exception when Cosmos DB HTTP response header is larger than 8192 bytes: https://github.com/Azure/azure-sdk-for-java/issues/6069
Vulnerability through dependency in the SDK v3.4.0: https://github.com/Azure/azure-sdk-for-java/issues/6433
CosmosSyncScripts null pointer exception in azure-cosmos: https://github.com/Azure/azure-sdk-for-java/issues/6281
Default consistency level parsing for Bounded Staleness and Consistent Prefix: https://github.com/Azure/azure-sdk-for-java/issues/6707
Null Value Holder change: https://github.com/Azure/azure-sdk-for-java/issues/6307
Closing this PR as there are breaking changes with azure-data-sdk-parent v1.2.0 with latest azure-cosmos pom files.
| gharchive/pull-request | 2019-12-13T01:48:52 | 2025-04-01T06:36:45.290080 | {
"authors": [
"kushagraThapar"
],
"repo": "Azure/azure-sdk-for-java",
"url": "https://github.com/Azure/azure-sdk-for-java/pull/6818",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
824133234 | [Cosmos] When are we getting Deno drivers?
Are Cosmos drivers for Deno in the works? We now have MongoDB https://github.com/denodrivers/deno_mongo and all the others -- but no Cosmos. Why don't I see any mention of this? Where are the drivers to use Azure blob storage? How can one use Deno within Azure without any hooks into the services layer?
#13281 is the issue that we are using to track all discussions around deno support in Azure SDKs
@zfoster, @southpolesteve If you have nothing to add here that will be specific to cosmos, I would recommend moving this conversation to #13281
cc @bterlson, @xirzec, @chradek
Yep that sounds good, thanks! nothing I can think of so closing this one
| gharchive/issue | 2021-03-08T03:37:18 | 2025-04-01T06:36:45.292495 | {
"authors": [
"aadamsx",
"ramya-rao-a",
"zfoster"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/issues/14157",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1101775443 | Sorting arm-consumption 'usage details' by resource group
I am using the arm-consumption package to get the usage details for a subscription and finding it difficult to sort the results in a meaningful way. I need to get them by resource group, but 1) it does not appear to be possible to filter the details by resource group when getting the data from azure and 2) the 'resource group' property on the usage details items does not correspond directly to the resource group, i.e., if the resource group is called 'ExampleGroup' (as per arm-resources or portal.azure.com), the usage details property 'resource group' is a longer string which appears to begin with the first few letters of the resource group but not in a consistent enough way to use to search. (There are a larger number of 'resource groups' appearing in arm-consumption than in arm-resources, and their names do not appear to correlate in a consistent, searchable way).
I would like to either be able to use a resource group filter when fetching the usage details, and/or have a property in the usage details that gives the resource group of the item as defined in arm-resources.
I have tried getting all the usage details data and sorting it myself by resource group, however then I encounter the problem where the resource group name that I need to sort by is not a property of the usage details items.
I have also tried looking to see if there is any existing API that provides a map between the arm-resources resource groups names and the arm-consumption resource group names, but have not so far found one.
@BasiaMH I think you can pass a scope parameter to get all the usage details within one resource group https://github.com/Azure/azure-rest-api-specs/blob/main/specification/consumption/resource-manager/Microsoft.Consumption/stable/2021-10-01/examples/UsageDetailsListByManagementGroup.json#L6 . let me know if that works for you ? Thanks
Thanks,
I can pass subscriptions/{subscriptionId} in the scope, which returns all
usage details for our subscription, but if I try to pass the
resourceGroups/{resourceGroupName} I get no results.
What is returned is an array with one item, containing a url
https://costmanagement.trafficmanager.net/subscriptions/c9e88f4c-b9f3-4941-b51e-c9cbbfb9742e/resourcegroups/dicomweb-group/providers/Microsoft.Consumption/usagedetails/Export
And when I navigate to the url it just gives an error message:
Either: "The resource you are looking for has been removed, had its name
changed, or is temporarily unavailable."
or:
{"message":"No HTTP resource was found that matches the request URI
'https://costmanagement.trafficmanager.net/subscriptions/c9e88f4c-b9f3-4941-b51e-c9cbbfb9742e/resourcegroups/dicomweb-group/providers/Microsoft.Consumption/usagedetails/Export'.","messageDetail":"No
route providing a controller name was found to match request URI
'https://costmanagement.trafficmanager.net/subscriptions/c9e88f4c-b9f3-4941-b51e-c9cbbfb9742e/resourcegroups/dicomweb-group/providers/Microsoft.Consumption/usagedetails/Export'"}
Basia
On Fri, Jan 14, 2022 at 3:22 AM Qiaoqiao Zhang @.***>
wrote:
@BasiaMH https://github.com/BasiaMH I think you can pass a scope
parameter to get all the usage details within one resource group
https://github.com/Azure/azure-rest-api-specs/blob/main/specification/consumption/resource-manager/Microsoft.Consumption/stable/2021-10-01/examples/UsageDetailsListByManagementGroup.json#L6
. let me know if that works for you ? Thanks
—
Reply to this email directly, view it on GitHub
https://github.com/Azure/azure-sdk-for-js/issues/19835#issuecomment-1012906456,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AFJXFDSC4B3G53GBYR3R4MDUV7MLNANCNFSM5L343HJQ
.
Triage notifications on the go with GitHub Mobile for iOS
https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675
or Android
https://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub.
You are receiving this because you were mentioned.Message ID:
@.***>
@BasiaMH Did you pass the scope like /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName} ?
That's exactly what I tried, yes.
(Before that I tried including resourceGroups in a filter, but that just
returned all the data unfiltered.)
On Sun., Jan. 16, 2022, 8:52 p.m. Qiaoqiao Zhang, @.***>
wrote:
@BasiaMH https://github.com/BasiaMH Did you pass the scope like
/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName} ?
—
Reply to this email directly, view it on GitHub
https://github.com/Azure/azure-sdk-for-js/issues/19835#issuecomment-1014065383,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AFJXFDWW4TZVRKOVU5UWZFLUWNY4DANCNFSM5L343HJQ
.
Triage notifications on the go with GitHub Mobile for iOS
https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675
or Android
https://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub.
You are receiving this because you were mentioned.Message ID:
@.***>
Looks like we need some help from service team.
@mitagarg Is it possible that you could help here, as I see you are the author of this PR https://github.com/Azure/azure-rest-api-specs/pull/17013
Hi, is there any update or further information on this problem? Or a different workaround someone could suggest to somehow get the usage and cost per resource or per resource group?
Thanks
@BasiaMH Hi, Since this is probably an issue on the service side, Could you open a ticket here https://docs.microsoft.com/en-us/answers/topics/azure-cost-management.html so that you can get direct support from the service team ?
| gharchive/issue | 2022-01-13T13:54:12 | 2025-04-01T06:36:45.310496 | {
"authors": [
"BasiaMH",
"qiaozha"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/issues/19835",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
439417810 | [Event Hubs] Re-visit all parameter validations
Relevant sections from General guidelines:
The client library will have a client object with several methods that call methods on the service. Service parameters are directly passed across the wire to an Azure service. Client parameters are not passed directly to the service, but used within the SDK to fulfill the request. Examples of client parameters include values that are used to construct a URI, or a file that needs to be uploaded to storage.
✅ DO validate client parameters.
⛔️ DO NOT validate service parameters. This includes null checks, empty strings, and other common validating conditions. Let the service validate any request parameters.
✅ DO validate the developer experience when the service parameters are invalid to ensure appropriate error messages are generated by the service. If the developer experience is compromised due to service-side error messages, work with the service team to correct prior to release.
Relevant sections from Typescript guidelines:
YOU SHOULD coerce incorrect types into an appropriate type, if possible. JavaScript users expect some amount of fuzziness with parameters as the standard library tends to coerce types if possible. TypeScript users should get pedantic types as they have opted in to types and expect errors.
Relevant work done in Service Bus regarding this: https://github.com/Azure/azure-sdk-for-js/issues/1145#issuecomment-481893308
Please use the parameter validations done in Service Bus as a reference
Main points to be kept in mind for parameter validations
Missing mandatory arguments should throw an error.
If expected type for an argument is string, use String() to convert given value to string. This is to ensure that we do our best to coerce given input to string for our Javascript users
If expected type for an argument is an array, given value is not an array, then make an array out of the given value
If given value doesnt match expected type and this results in us not being able to form a request to the service, then throw error.
For reference, see what was done for Service Bus
Done with #3652
| gharchive/issue | 2019-05-02T03:55:14 | 2025-04-01T06:36:45.316493 | {
"authors": [
"ramya-rao-a"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/issues/2660",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
608757284 | Propagate changes to work around DNS resolution issue in Linux
See: https://github.com/Azure/azure-sdk-for-java/pull/10576/files
We've seen issues where DNS resolution fails in Ubuntu 18. This workaround fixes that problem. This only needs to run on Ubuntu 18 instances.
@chradek, @richardpark-msft Any chance this can solve our dns woes with linux in SB and EH?
Closing this issue. We already have this integrated with a step to bypass local DNS using common template.
| gharchive/issue | 2020-04-29T03:47:20 | 2025-04-01T06:36:45.318443 | {
"authors": [
"danieljurek",
"praveenkuttappan",
"ramya-rao-a"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/issues/8594",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
813997040 | Bertong sms impl
This draft PR is to solicit feedback, primarily on whether what i'm testing is appropriate and consistent with what we hve done
/azp run python - communication - tests
/azp run js - communication - tests
/check-enforcer reset
/azp run js - communication-sms - tests
/azp run js - communication-sms - tests
/azp run js - communication-sms - tests
/azp run js - communication-sms - tests
/azp run js - communication-sms - tests
/azp run js - communication-sms - tests
| gharchive/pull-request | 2021-02-23T00:55:27 | 2025-04-01T06:36:45.321170 | {
"authors": [
"JoshuaLai",
"beltr0n"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/pull/13927",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2158349036 | push communication-email recordings
fix https://github.com/Azure/azure-sdk-for-js/issues/28656
API change check
API changes are not detected in this pull request.
| gharchive/pull-request | 2024-02-28T08:04:40 | 2025-04-01T06:36:45.322550 | {
"authors": [
"azure-sdk",
"kazrael2119"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/pull/28697",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2617358821 | Post release automated changes for databoundaries releases
Post release automated changes for azure-arm-databoundaries
API change check
API changes are not detected in this pull request.
| gharchive/pull-request | 2024-10-28T05:40:53 | 2025-04-01T06:36:45.324009 | {
"authors": [
"azure-sdk"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/pull/31544",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
561731350 | [core-amqp] Fixed a build error on master
Seems like a fitting solution for a problem we're having on master.
The error is visible here: https://dev.azure.com/azure-sdk/public/_build/results?buildId=255040&view=logs&j=23e2e6de-b7c3-5918-7121-f16b46172e49&t=9c960e92-0110-52f6-5b54-96f3b7fea4ea&l=697
This isn't necessary. My master was outdated.
| gharchive/pull-request | 2020-02-07T16:09:24 | 2025-04-01T06:36:45.325593 | {
"authors": [
"sadasant"
],
"repo": "Azure/azure-sdk-for-js",
"url": "https://github.com/Azure/azure-sdk-for-js/pull/7290",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
581920916 | [BUG] Memory Leaking When SDK Combined With Application Insights
Describe the bug
When using the Azure.Storage.Queues (v 12.3.0) NuGet package in combination with Microsoft.ApplicationInsights.AspNetCore (v 2.13.1) in a .Net core 3.1 application that is constantly polling queues for new messages, we find that the number of objects built up in memory is steadily increasing, specifically:
Microsoft.ApplicationInsights.DependencyCollector.Implementation.AzureSdkDiagnosticsEventHandler - creating about 5000 objects a minute and the garbage collector is not disposing them.
Expected behavior
That the AI objects would be cleaned up by the GC
Actual behavior (include Exception or Stack Trace)
Eventually the app will start throwing OutOfMemoryExceptions.
To Reproduce
Have the following in your app startup:
public void ConfigureServices(IServiceCollection services)
{
services.AddApplicationInsightsTelemetry();
}
Add a hosted service:
var host = new WebHostBuilder()
.UseKestrel()
.ConfigureServices(c => c.AddHostedService<QueueWorker>())
Have the worker constantly new up a QueueClient to check for queue messages.
Environment:
Azure.Storage.Queues 12.3.0
.NET runtime version 3.1.101
IDE and version : Visual Studio Professional v 16.4.5
For step 3, why not have the worker re-use the same QueueClient? The clients are thread-safe.
That has proven to be the workaround we have gone with - storing any queue clients that we have created in a static dictionary that remains for the lifetime of the app (we have a web app that uses this shared code and would be newing up the clients pretty heavily to create queue messages).
I will leave it open and for others to decide if this is a bug that needs fixing or if it was simply of a case of us "using it wrong" in which case it can be closed.
We're tracking getting this documented for all clients in this issue. https://github.com/Azure/azure-sdk-for-net/issues/8941.
Hi there
We understand the recommended way is to use a single client though isn't still a bug having a memory leak because many clients are created?
In our specific case, we created a lot of clients for an hour or so at night and the memory is never release even days after.
We will use a single client in the future, so thanks to @speedy-ms
Thanks
Phil
| gharchive/issue | 2020-03-16T01:35:30 | 2025-04-01T06:36:45.333068 | {
"authors": [
"Petermarcu",
"jesuissur",
"seanmcc-msft",
"speedy-ms"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/issues/10626",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1631585358 | [QUERY] Azure .NET SDK for web app/function app creation and deployment
Library name and version
Microsoft.Azure.Management.AppService
Query/Question
Hi,
I would like to do the below commands from the CLI in the SDK instead.
$'az functionapp create -p {_appPlanName} --name {_globalFunctionName} --resource-group {_resourceGroupName} --runtime dotnet --storage-account {_storageAccountName} --functions-version 4'
$'az functionapp deployment source config-zip -g {_resourceGroupName} -n {_globalFunctionName} --src {_zipName}'
Looks like there's SDK Microsoft.Azure.Management.AppService could create and deploy web app, but there's no example for this. Could you please share some example for C# SDK creation of Azure function, SDK deploy of Azure function and SDK deploy of web app
Thanks
Environment
No response
function app is one kind of web site.
Here is an example of web site creation. Need to set Kind as functionapp to create a function app, and you might experiment more if you can fit the settings for function app into WebSiteData
I did not find any SDK regarding the deployment though.
It requires details of the service itself, so it would be best to involve the service team to provide a sample of the usage.
@jsquire Could you please help route this issue to service team for further assistance?
@live1206 : Service teams do not generally provide samples based on our SDKs. This is something that our team would own and coordinate with the service team directly, if needed. Please work with @ArthurMa1978 to determine next steps.
Hi @Arvinorange. Unfortunately, this repository is focused on the Azure SDK for .NET, service teams do not monitor these issues. To involve the service team, your best path forward would be to open an Azure support request.
@live1206 : Service teams do not generally provide samples based on our SDKs. This is something that our team would own and coordinate with the service team directly, if needed. Please work with @ArthurMa1978 to determine next steps.
@jsquire AFAIK, The sample of how to use SDK for a specific scenario is auto-generated based on examples in Spec repo.
The service team need to provide us a detailed example of the request and response how to create a function app.
If we take a look at the example of CreateOrUpdateStaticSiteBuildFunctionAppSettings, it generated a sample CreateOrUpdateAppSettings_CreatesOrUpdatesTheFunctionAppSettingsOfAStaticSiteBuild, which does not provide any details regarding how to create a function app with .NET SDK.
I have checked with @ArthurMa1978 we can't do anything here if the service team doesn't provide the scenario example with details.
| gharchive/issue | 2023-03-20T08:02:05 | 2025-04-01T06:36:45.342209 | {
"authors": [
"Arvinorange",
"jsquire",
"live1206"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/issues/35002",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
480393080 | [BUG] Azure.Core.ResponseOfT contains disposable Response
Describe the bug
Response<T> contains a disposable Response property. There is no way to dispose of the inner Response object. This is a memory leak and will cause additional overhead for garbage collection.
Information Checklist
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
[x] Bug Description Added
[x] Repro Steps Added
[x] Setup information Added
@pakrym this is related to the recent PR
Not disposing a reference doesn't cause a memory leak as long as it doesn't hold any common/unmanaged resources without having a finalizer.
In the case of the Response the only part that holds onto shared resource is the content stream. It also depends if response buffering is enabled:
In the case where buffering is enabled the process of buffering itself would dispose the underlying stream and avoid resource leak, buffered stream is just a memory stream and can be collected.
In the case where buffering is disabled, it's clients responsibility to either dispose Response it got from the pipeline or return Response where T is IDisposable and consumer of the client SDK has to dispose the value.
In addition not disposing something doesn't add any GC overhead because no matter if an object is disposed GC still has to walk the graph and collect it.
Response ContentStream is a Stream. There is no guarantee that it is a MemoryStream. Does Response need to be updated to be MemoryStream instead of Stream?
How will users know that there is inner property that needs to be disposed of?
Would a better solution be to use Memory or Span instead of Stream? It would remove all the dispose logic and it would give better performance.
The response should be able to handle both buffered and streamed responses, Memory or MemoryStream would only work for buffered responses.
We know that Response.ContentStream would be MemoryStream for buffered responses because we control the buffering logic and create the stream.
Any reason not to make ChannelEnumerableSubscription itself IAsyncEnumerator ?
My issues with this is it is dependent on a implementation detail, and not the contract. If someone changes that implementation in the future then this will be broken. There is also no way for users to know if they need to dispose of the inner stream or not to dispose of it.
What scenario requires supporting stream? Would users actually get better use out of always having a buffered Memory object?
My issues with this is it is dependent on a implementation detail, and not the contract. If someone changes that implementation in the future then this will be broken. There is also no way for users to know if they need to dispose of the inner stream or not to dispose of it.
Users never have to dispose the inner stream. Users only need to dispose something when the return type is Response then they have to dispose the response.Value.
What scenario requires supporting stream? Would users actually get better use out of always having a buffered Memory object?
Downloading large files from blob storage, for example.
Users never have to dispose the inner stream. Users only need to dispose something when the return type is Response then they have to dispose the response.Value.
I'm a little confused here. Response<T> has an inner Response object that is disposable. If Response contains a network or some other Stream type that needs to be disposed of how is it getting disposed of?
I'm a little confused here. Response has an inner Response object that is disposable. If Response contains a network stream or some other stream type that needs to be disposed of how is it getting disposed of?
We are controlling the clients and the pipeline and wouldn't return Response that has a network stream in it.
| gharchive/issue | 2019-08-13T22:11:14 | 2025-04-01T06:36:45.352585 | {
"authors": [
"j82w",
"pakrym"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/issues/7286",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
674646306 | Enable a way to Unregister Message Handler and Session Handler
Currently our SDK doesn't support a way to unregister message handler and session handler so customers are getting ObjectDisposedException exceptions when receivers are closed due to connection closed caused by regular application upgrades etc. Exposing this functionality to unregister handler and await for pending receives and message handling operations to finish to allow for a graceful unregister on the handler they previously registered. Customer are expected to call this function before they close down their handler. Customer can register again after unregister as they are independent operations.
Is there an issue that discusses this change? Given this is a new feature I'd expect to see some transparency about it.
/cc @JoshLove-msft @jsquire
@SeanFeldman Hi Sean! Thanks for your comments. This is only a draft PR. This is a feature request from Azure Functions team. I will create an issue to discuss this to allow more transparency in design.
I have validated a version of these changes (commit "fbd31d0") with Azure Functions Service Bus extension and did not find any issues. Tested the following methods:
QueueClient.UnregisterSessionHandlerAsync(TimeSpan)
SubscriptionClient.UnregisterSessionHandlerAsync(TimeSpan)
3 MessageReceiver.UnregisterMessageHandlerAsync(TimeSpan)
I also attempted to call SubscriptionClient.RegisterSessionHandler() while SubscriptionClient.UnregisterSessionHandlerAsync(TimeSpan) was processing and got an exception as expected. Calling RegisterSessionHandler() after UnregisterSessionHandlerAsync(TimeSpan) completed worked normally as expected.
Nice work @DorothySun216! Approving from my side.
used version: 4.2.1
I want to use UnregisterMessageHandler to "pause" the ISubscriptionClient. aka stop processing messages from the Service Bus for a while. The only problem I encounter, is that after I unregister and try to re-register a handler, the runningTaskCancellationTokenSource in MessageReceiver.cs does not get a new instance, but keeps using the disposed one. That is why I get the following exception:
System.ObjectDisposedException: The CancellationTokenSource has been disposed.
Suggestion to also re-instantiate the runningTaskCancellationTokenSource just like the receivePumpCancellationTokenSource in MessageReceiver.cs (line 1333).
used version: 4.2.1
I want to use UnregisterMessageHandler to "pause" the ISubscriptionClient. aka stop processing messages from the Service Bus for a while. The only problem I encounter, is that after I unregister and try to re-register a handler, the runningTaskCancellationTokenSource in MessageReceiver.cs does not get a new instance, but keeps using the disposed one. That is why I get the following exception:
System.ObjectDisposedException: The CancellationTokenSource has been disposed.
Suggestion to also re-instantiate the runningTaskCancellationTokenSource just like the receivePumpCancellationTokenSource in MessageReceiver.cs (line 1333).
@RicoBakels thanks for reaching out. Are you using our service bus SDK directly or using Azure Functions?
@RicoBakels thanks for reaching out. Are you using our service bus SDK directly or using Azure Functions?
@DorothySun216 I am using the service bus SDK
@DorothySun216 I am using the service bus SDK
@RicoBakels I see. I will work on a repro. So you just register a handler and then unregister and then register again, and during this time the exception occurred? Can you show a code snippet for your use case if it is convenient?
@RicoBakels I see. I will work on a repro. So you just register a handler and then unregister and then register again, and during this time the exception occurred? Can you show a code snippet for your use case if it is convenient?
@DorothySun216 Thanks for the fast reply. Here is a code snipped where I use the ISubscriptionClient in my class.
At startup of the (Service Fabric) service, The message handler will get registered with the OpenAsync method. When for some reason the service needs to stop handling messages from the Service Bus for a while, the service needs to pause the handling of messages by calling the Pause method.
Then when the service needs to start handling messages again, I call the OpenAsync method again to register the handler.
// (re-)register message handler.
public Task<string> OpenAsync(CancellationToken cancellationToken)
{
_logger.LogInformation("AzureServiceBusReaderService starting");
var messageHandlerOptions = new MessageHandlerOptions(ExceptionReceivedHandlerAsync)
{
MaxConcurrentCalls = 1,
AutoComplete = false
};
_subscriptionClient.RegisterMessageHandler(ProcessMessagesAsync, messageHandlerOptions);
return Task.FromResult(_subscriptionClient.SubscriptionName);
}
// Unregister message handler so messages won't get picket up from Service Bus.
public void Pause() => _subscriptionClient.UnregisterMessageHandlerAsync(TimeSpan.FromSeconds(1));
@DorothySun216 Thanks for the fast reply. Here is a code snipped where I use the ISubscriptionClient in my class.
At startup of the (Service Fabric) service, The message handler will get registered with the OpenAsync method. When for some reason the service needs to stop handling messages from the Service Bus for a while, the service needs to pause the handling of messages by calling the Pause method.
Then when the service needs to start handling messages again, I call the OpenAsync method again to register the handler.
// (re-)register message handler.
public Task<string> OpenAsync(CancellationToken cancellationToken)
{
_logger.LogInformation("AzureServiceBusReaderService starting");
var messageHandlerOptions = new MessageHandlerOptions(ExceptionReceivedHandlerAsync)
{
MaxConcurrentCalls = 1,
AutoComplete = false
};
_subscriptionClient.RegisterMessageHandler(ProcessMessagesAsync, messageHandlerOptions);
return Task.FromResult(_subscriptionClient.SubscriptionName);
}
// Unregister message handler so messages won't get picket up from Service Bus.
public void Pause() => _subscriptionClient.UnregisterMessageHandlerAsync(TimeSpan.FromSeconds(1));
| gharchive/pull-request | 2020-08-06T22:54:54 | 2025-04-01T06:36:45.363122 | {
"authors": [
"DorothySun216",
"RicoBakels",
"SeanFeldman",
"sidkri"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/pull/14021",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1768431109 | Created wrapper class which contains information to rehydrate
Moved around dependencies, job plan models to create Checkpointer internally as well.
API change check
APIView has identified API level changes in this PR and created following API reviews.
Azure.Storage.DataMovement
| gharchive/pull-request | 2023-06-21T20:33:07 | 2025-04-01T06:36:45.365341 | {
"authors": [
"amnguye",
"azure-sdk"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/pull/37153",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1991741851 | Increment version for storage releases
Increment package version after release of Azure.Storage.Common
API change check
API changes are not detected in this pull request.
| gharchive/pull-request | 2023-11-14T00:51:38 | 2025-04-01T06:36:45.366427 | {
"authors": [
"azure-sdk"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/pull/39967",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2182714546 | [ACS][CallAutomation]Move SourceCallerIdNumber from Answer to Transfer api
Contributing to the Azure SDK
Please see our CONTRIBUTING.md if you are not familiar with contributing to this repository or have questions.
For specific information about pull request etiquette and best practices, see this section.
Run autorest to updated a lot models include media teams work
In CallMedia.cs, StartHoldMusicAsync and StartHoldMusic is updated to pass the build and left comment for media team to work on the StartHoldMusicOptions.Loop and operationCallbackUri
Move SourceCallerIdNumber from Answer to Transfer api
API change check
APIView has identified API level changes in this PR and created following API reviews.
Azure.Communication.CallAutomation
| gharchive/pull-request | 2024-03-12T21:30:15 | 2025-04-01T06:36:45.369962 | {
"authors": [
"azure-sdk",
"fangchen0601"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/pull/42642",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2334798863 | [HDInsight on AKS] Api version 2024-05-01 support
Contributing to the Azure SDK
Support api version 2024-05-01.
Preview to Stable.
Please see our CONTRIBUTING.md if you are not familiar with contributing to this repository or have questions.
For specific information about pull request etiquette and best practices, see this section.
API change check
APIView has identified API level changes in this PR and created following API reviews.
Azure.ResourceManager.HDInsight.Containers
The GA plan has changed. Stop this thread. Thanks.
| gharchive/pull-request | 2024-06-05T03:03:22 | 2025-04-01T06:36:45.372859 | {
"authors": [
"azure-sdk",
"yuc-Li"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/pull/44393",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2539821981 | AZD-CM integration
Added the ability to provision CDK CM infrastructure using azd.
Steps to try it:
Execute the CloudMachineTests.Configure (this will create bicep files from the CM CDK)
In the test's bin folder, run azd init
pick "Minimal" template
you can now do azd provision
API change check
APIView has identified API level changes in this PR and created following API reviews.
Azure.Provisioning.CloudMachine
| gharchive/pull-request | 2024-09-20T23:45:43 | 2025-04-01T06:36:45.375262 | {
"authors": [
"KrzysztofCwalina",
"azure-sdk"
],
"repo": "Azure/azure-sdk-for-net",
"url": "https://github.com/Azure/azure-sdk-for-net/pull/46106",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
580757152 | [event hubs] support receiver redirect
Background
The Event Hubs service supports redirects for amqp receiver links. When the client creates an amqp receiver link with redirects enabled, the service will respond with an amqp:link:redirect error if it also supports redirects. The error contains the information needed to create a new connection to the service.
The main benefit of receiver redirect is it allows a more direct connection to the host actually sending events to the client which should reduce latency.
API changes
The Event Hub clients responsible for consuming events (e.g. EventHubConsumerClient) should accept a new option to enable redirect that has a default value of false.
Proposal:
{ enableRedirect: true } // new option 'enableRedirect'
Reasoning
It is important to make this option opt-in for 2 reasons:
Currently 1 client creates 1 connection. This changes this assumption to 1 client can create many connections. This has implications since there is a limit to the number of connections that can be open at one time to an Event hub.
By default the SDK speaks AMQP over ports 5671/5672. With redirect enabled, new connections may be made on ports 104xx. Users may need to update their firewall rules to support redirect:
https://docs.microsoft.com/en-us/azure/service-bus-messaging/service-bus-amqp-protocol-guide#amqp-outbound-port-requirements
Details
Redirect walkthrough
User instantiates a consumer client with redirect enabled.
User makes call to receive events (this causes a receiver link to be created using customer-provided credential/host info).
Connection/cbs session/receiver link are created.
The receiver link must have amqp:link:redirect included in the list of desired_capabilities if redirect is enabled.
This must be omitted if redirect is not enabled.
The receiver link recieves a LinkRedirectError.
The SDK parses the hostname/port/address info from the error.
A new amqp connection is created using the hostname/port from the error.
When the cbs session is created, the applicationProperties.name value should be the address extracted from the LinkRedirectError.
The SDK creates the receiver link on the new connection and begins receiving events.
Considerations
It is possible for every receiver link to have different hostname/port/address values from one another in the LinkRedirectError. You may create a new connection for every LinkRedirectError. Optionally, you may use a connection for multiple receiver links if the links share a hostname/port.
Creating a new connection for each redirect is simpler to implement because it is easier to know when the connection should be closed, but does potentially increase connection density.
When to re-use the original connection?
It is possible that the server node used for receiving events can change. When the SDK sees an amqp:connection:forced error from the service, or the connection is closed, then the original connection should be used to create a new receiver link. At this point a new LinkRedirectError should be received and the redirect walkthrough can be followed.
Otherwise, for transient issues the connection created for the link after a LinkRedirectError can be re-used.
Reference
There is a draft PR for the JavaScript SDK that has the basics implemented for this feature.
https://github.com/Azure/azure-sdk-for-js/pull/7782
goal for April: design
| gharchive/issue | 2020-03-13T17:56:48 | 2025-04-01T06:36:45.386272 | {
"authors": [
"chradek",
"yunhaoling"
],
"repo": "Azure/azure-sdk-for-python",
"url": "https://github.com/Azure/azure-sdk-for-python/issues/10296",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
657715748 | azure.storage.blob.BlobClient.from_blob_url() parses incorrect blob name and container name
Package Name: azure-storage-blob
Package Version: 12.3.2
Operating System: Windows 10 Version 10.0.19041 Build 19041, running WSL 1
Python Version: 3.8.3
Describe the bug
azure.storage.blob.BlobClient.get_blob_properties() returns incorrect "name" and "container" properties
To Reproduce
Steps to reproduce the behavior:
Code
from azure.storage.blob import BlobClient
PUBLIC_BLOB_URI = 'https://lilablobssc.blob.core.windows.net/nacti-unzipped/part0/sub000/2010_Unit150_Ivan097_img0003.jpg'
blob_client = BlobClient.from_blob_url(PUBLIC_BLOB_URI)
properties = blob_client.get_blob_properties()
print(properties['name'])
print(properties['container'])
Running Code in Python REPL:
>>> from azure.storage.blob import BlobClient
>>>
>>> PUBLIC_BLOB_URI = 'https://lilablobssc.blob.core.windows.net/nacti-unzipped/part0/sub000/2010_Unit150_Ivan097_img0003.jpg'
>>>
>>> blob_client = BlobClient.from_blob_url(PUBLIC_BLOB_URI)
>>> properties = blob_client.get_blob_properties()
>>> print(properties['name'])
2010_Unit150_Ivan097_img0003.jpg
>>> print(properties['container'])
sub000
Expected behavior
A clear and concise description of what you expected to happen.
The blob name should be "part0/sub000/2010_Unit150_Ivan097_img0003.jpg" instead of "2010_Unit150_Ivan097_img0003.jpg" and the container should be "nacti-unzipped" instead of "sub000".
Update: the issue is actually in azure.storage.blob.BlobClient.from_blob_url().
from azure.storage.blob import BlobClient
PUBLIC_BLOB_URI = 'https://lilablobssc.blob.core.windows.net/nacti-unzipped/part0/sub000/2010_Unit150_Ivan097_img0003.jpg'
blob_client = BlobClient.from_blob_url(PUBLIC_BLOB_URI)
print(blob_client.blob_name)
print(blob_client.container_name)
Running Code in Python REPL:
>>> from azure.storage.blob import BlobClient
>>>
>>> PUBLIC_BLOB_URI = 'https://lilablobssc.blob.core.windows.net/nacti-unzipped/part0/sub000/2010_Unit150_Ivan097_img0003.jpg'
>>>
>>> blob_client = BlobClient.from_blob_url(PUBLIC_BLOB_URI)
>>> print(blob_client.blob_name)
'2010_Unit150_Ivan097_img0003.jpg'
>>> print(blob_client.container_name)
'sub000'
Hi @chrisyeh96
Thanks for reporting this and post that pr. It's the SDKs problem with parsing blob url with sub directories! Sorry for the inconvenience.
the fix has been released, you can do pip install azure-storage-blob==12.4.0 to use it.
| gharchive/issue | 2020-07-15T23:05:37 | 2025-04-01T06:36:45.393079 | {
"authors": [
"chrisyeh96",
"xiafu-msft"
],
"repo": "Azure/azure-sdk-for-python",
"url": "https://github.com/Azure/azure-sdk-for-python/issues/12563",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
794491432 | [API Management Management] Impossible to create new API using sdk
Package Name: Azure management Api Management
Package Version: 1.0.0
Operating System: Windows
Python Version: 3.7
Describe the bug
It seems as the api management client has a bug when trying to create or update an API. The package tries to poll for a result but includes uneccessary data in the request which throws an error in the aiohttp package
To Reproduce
Steps to reproduce the behavior:
Create a client for interacting with API management
creds = DefaultAzureCredential()
client = ApiManagementClient(
credential=creds, subscription_id=settings.subscription_id
)
create a version set
api_version_set=client.api_version_set.create_or_update(
resource_group_name=settings.resource_group_name,
service_name=settings.api_management_name,
version_set_id="anystuff",
parameters=ApiVersionSetContract(
display_name="title",
versioning_scheme="Segment",
description="Version configuration",
),
)
And then create any form of API.
api=client.api.begin_create_or_update(
resource_group_name=settings.resource_group_name,
service_name=settings.api_management_name,
api_id="wss-mp-test-order",
parameters=get_parameters(oai, api_version_set.id),
)
This throws an error:
return _RequestContextManager(self._request(method, url, **kwargs))
TypeError: _request() got an unexpected keyword argument 'path_format_arguments'
which, as far as I can see, results from these lines in the sdk:
https://github.com/Azure/azure-sdk-for-python/blob/79a64bb9d40610d1696819a2595a49d18d7a7ad6/sdk/apimanagement/azure-mgmt-apimanagement/azure/mgmt/apimanagement/aio/operations/_api_operations.py#L420-L427
where the path_format_arguments are interpreted in the end as parameters to the GET request polling for a result.
Expected behavior
A created API and successful response in the code
Screenshots
N/A
Additional context
I had to upgrade from 0.2.0 because of a different set of bugs (not reported yet, but at least one still persist). Now everything broke, even the things that I had a workaround for before.
Hi @ludokriss, this version of SDK is using a preview-version of API.
It seems that something have changed in service side.
We will update it in the near future.
@00Kai0 any updates?
| gharchive/issue | 2021-01-26T19:14:31 | 2025-04-01T06:36:45.400303 | {
"authors": [
"00Kai0",
"ludokriss",
"xiangyan99"
],
"repo": "Azure/azure-sdk-for-python",
"url": "https://github.com/Azure/azure-sdk-for-python/issues/16356",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.