Hi Niclas Schmidt, in particular after you've attempted all the no-brainer solutions swapping out datasets, images, even resources and locations… come on, azure, work with us here
alright, let's break that down. first of all, that error "Training failed unexpectedly" is a bit ambiguous, but it generally appears if the service encounters a hitch that it didn't anticipate. it could be a backend glitch, or possibly there's some kind of weird data formatting issue even if it used to work. lets see what you can do to try to resolve it
check your image tags. at times, the custom vision service can become finicky if the tags are not uniform or are not present. ensure that each image has a tag, and the spelling of the tag is uniform throughout. microsoft's docs discuss that here.
resize those images. if ur uploads are enormous (like 4K+ resolution), the service could slow down. see if scaling them down to under 6MB makes a difference. the docs are recommending keeping them between 256x256 to 1024x1024 pixels here.
region does indeed matter. you mentioned you tested different locations, but make sure ur custom vision resource and ur storage (in case ur utilizing one of those) are in the same region. getting them mixed up can lead to silent fails.
- quota limits. Sneaky one if ur subscription crossed over into one of those obscure limits (e.g. api calls or training time), the service could just… give up. take a look at ur azure metrics to exclude that.
occasional glitch. azure is big, but it's not perfect. the service just burps from time to time. wait sixty minutes, retry, and pray to the cloud gods
So, hope that will help, at least i try to help
Best regards,
Alex
P.S. If my answer help to you, please Accept my answer
PPS That is my Answer and not a Comment
https://ctrlaltdel.blog/