~600 images with 4 seconds exposure without a tracker. This is my first ever (successful) try of any object on the night sky. I did a post a few weeks ago asking for a camera setup, and I went and bought a panasonic lumix g85 with the 45-150 f4-5.6 lens I had in mind before that. I know this is not the ideal setup, but I am still very proud of that image. Although I would appreciate any feedback on how to improve my skills.
For untracked this looks not too bad. To improve I would do the following:
- in Siril: remove green noise, then photometric colour calibration
- reduce exposure length to 1s
- work on background extraction (maybe try GraXpert)
- if you are not already doing: take calibration frames, especially flats
Are you interested in sharing the raw stacked file? I use a (paid) deconvolution tool called BlurXterminator and I wonder if it can handle such extreme star shapes. If it works I will of course send you the file.
Thanks for the feedback. You can try it with your software. Here is the file: https://drive.proton.me/urls/M8309JQDYR#rc9F6A9RYyCk
But I read some negatives online about how it uses hubble images to calibrate images. Here is a forum post: https://forum.astronomie.de/threads/blurxterminator-nein-danke.336225/ It is in german, but I’m sure you will understand it. Let me know what you think
Edit: Since the software is probably trained on images from better telescopes/observatories it is most likely using their images to calibrate mine. I have no direct problem with that, but I wouldn’t consider this to be my own anymore. Personally I don’t like using AI to enhance or fix some of my stuff.
But I’d still like to see the results if it works.
Thanks for sharing! The software worked better than expected on your image!
On my research before purchasing the program I also stumbled over your linked forum post. However I found it very misleading, as the software does not generate details learned from other images but only works with data already in your image. As it is a deconvolution tool results can deviate slightly from the true nature. But that has little to do with AI being used here. I needed a whole semester at university to truly understand the maths behind it. My biggest problem is that the software isn’t open source so one can’t look into all the details. But there are already people working on open alternatives.
But this already is a very specific problem, don’t forget that the biggest difference for good quality makes the data itself. I wish best of luck on your journey!
Oh and I forgot to mention that one other advice I would give is to search the darkest location to shoot from that you can access. Lightpollutionmap really helps with finding such places.
Edit: unstreched file with only background extraction and deconvolution: https://drive.proton.me/urls/QD4870ZMF4#uyVBYxKgWxgb