-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
where to get data for training #300
Comments
Ok I found this (I see the blog wasn't updated so there is a more recent download) hopefully that will get me the bulk of the street annotations (road, pavement) and some of the person data ,but I'd be waiting for unlock and approval of left/man, right/man, left/woman, right/woman to get the best of whats there ( I appreciate unlocking those images will take time .. if you could prioritise them based on the number of total annotations + inclusion of those specific labels (that give me orientation information) that would maximise the value per unlock) I can't get the python library working.. what would be ideal is a plain download script without any other dependencies (docker, tensorflow) .. I remember we had a python program to upload images with custom labels which was great but I can't find that again (its on a broken laptop). Essentially I'm after the inverse of that now I can try to dig around but if you can make a way to just get the raw images and outlines (json dumps of the type the site shows in explore, image files named with their ID?) without any other dependencies/installs that will help me out I've gone down the PyTorch route; I'm needing to make a custom net with the ability to hint multitask learning based on as much as is available from as many images as possible: [1] baseline , no annotations is an autoencoder (eg use all the images) |
Sorry for being a bit unresponsive lately - renovating my flat takes unfortunately takes away most of my free time :/ Awesome to hear that you got back to experimenting with neural nets - that sounds really interesting!
I think there's nothing like that right now, unfortunately. But I'll try to hack a small Python script together for you. What are you interested in exactly? The images together with the labels or do you want to also have the annotations? Regarding the labels: I plan to unlock/approve a bunch of those |
right good luck with the move ! I appreciate that will chew up a lot of time and focus.
Images, Label List (including un-annotated)+ Annotations (polygons, bounding boxes): The format I saw it show in "explore -> export" through the UI is fine (it seems disabled at the moment, I guess things might be in flux). Because I know my way around the dataset.. I hope I'll be able to setup multi-task training that makes the most of the different permutations we have.. and the graph information There's someone else interested in collaborating on making nets to enhance retro games. Lots of pieces to get this working. I've written something in openCL to run conv-nets that I can integrate in-engine (thats to get going, eventually I'll need metal compute, CUDA etc), but I will stick with PyTorch as the main workhorse for actual training. I'm also hoping to use StableDiffusion to enhance the images 'somehow' (completely open). This net is awesome BUT pretty heavyweight (5gb? of weights). I figure we can make lightweight nets that can run in game , and the open-ended image-monkey data will focus things beyond a plain autoencoder as a starting point You've probably heard of the controversy around Stable Diffusion - i.e. some people think the legality of their source data is dubious, and like YouTube eventually the issue will come to ahead. Seeing Stable Diffusion going, and seeing the end of the GPU shortage has got my enthusiasm up.. all that voluntary annotating will not be in vain. And I'm also hoping general interest in these kind of use cases will get more people interested in contributing. I also want to look into adapting my engine to render additional channels to feed the NN's. (again pre-process textures to create an embedding, essentially screen-space texturing) I dont have photorealism in mind.. more like "playing a comic book / concept art", something useable for indie/retro games, rather than the AAA look. |
https://github.com/dobkeratops/convnet_stuff very early days but here's the repo I'm sharing my current experiments in. I want to write an imagemonkey data loader there. |
If you do get time to enable these - maybe you could also add "clothing", there's 1000 examples annotated ; it's useful to be precise in the difference between the person's outline and any coats, dresses etc , + surface texture, and it gives information that wouldn't be inferable from 3d scans (these polys can also be refined with the material system) |
short update: I also added a small Python snippet which demonstrates how to query the database and export data: https://github.com/ImageMonkey/imagemonkey-libs/blob/master/python/snippets/export.py. The script requires the |
awesome, that looks really interesting! |
https://github.com/dobkeratops/convnet_stuff , the readme explains a bit more now I've set it up for multiple input or output images , I want this for graphics use cases (eg color + normal maps.. video frames etc). In this test I split the RGB into 2 images, and gave it some features painted out in the output, and it's learned to reconstruct, and which bits to fill in probably overfitted badly though, this is just a quick test that runs for a few mins with a dozen images. it's a "u-net", encoder-decoder with skip-connections. |
Running this now.. so far so good, I'm able to get a bunch of images and write JSON for the annotations. |
download/render code here. road/pavement, blending here assumes mutually exclusive annotations.. (not true in general but I might be able to improve what this looks like with hard coded label priority.. some are more likely background vs foreground features) all person part annotations , preview tries to show overlapping annotations better |
Would there be a way to use these ID's to open an image in unified? I can easily make something here to render a page with clickable links, although updating the thumbnails might be trickier. It's really easy to flick through these thumbnails to find images that need more work - and to find the best ones to train on |
Awesome progress - thanks for sharing!
This should work:
There are also endpoints to label & annotate images. But especially the annotate endpoint is a bit cumbersome to use (mostly because the unified mode was added later on and the schema of the database and the API endpoints weren't fitting the unified mode nicely. So I had to work around some limitations). But I think the label endpoint(s) aren't that complicated to use. So in case you want to give it a try and add it to your tool, I could look the API calls up. |
works fine, thats great |
got that working now- it spits out an html page of thumbnails, they can be clicked, and it fires up unified editor.. awesome! |
generated along with the training images maybe you could integrate these thumbnail summaries somewhere in the site (even if they dont update when you annotate.. I think you had concerns about how long it takes the server to do this sort of thing) I might experiment a bit with some static layouts for browsing all the images this way |
trying to get data for the road/pavement task - with these thumbnail previews I can find the images with gaps and fill them . |
hi again,
I have started using PyTorch with my RTX3080, which seems fast enough to experiment with (I'd been put off in the past by training times)
I'm tinkering with denoising autoencoders (and eventually want to try using parts of the stable diffusion model thats all the rage now, but initially I'm experimenting with my own smaller examples); currently my intention is to make something to enhance lo-res retro/indie game art with neural nets, so I'm getting the pieces in place.. a simple runtime, and something in PyTorch for training
what I've got in mind is
I'd like to grab the imagemonkey database , to setup training for the following in particular:
Road+Pavement
left/man , right/man. left/woman,right/woman.
.. plus the entire label list
these labels and a lot of the examples are locked.. would you be able to approve them and make it accessible to be somehow
There's 3000 person outlines in the format I'm after (1500 x left/right man , 1500 x left/right woman)
I'll try to setup multi-task training -
I've managed to adapt my own "DataLoader" in PyTorch for a denoising autoencoder, which I'll extend to do all this
(the repo has a simple OpenCL inference test thats intended to grow into a little library to integrate with game engines, and my PyTorch training setup)
The text was updated successfully, but these errors were encountered: