April 9, 2016 at 8:46 pm #102111
First off, three cheers for making the AR Sandbox tech so accessible. I made one for my office with very little difficulty thanks to your excellent software and instructions.
I work with Resource Management Associates in Davis, CA, a small engineering firm specializing in numerical modeling for water resources. I would like to use the sandbox to illustrate and physically play around with concepts related to a wide range of water issues. As is, I have some fun illustrations of reservoir management. We do a lot of analysis for tidal marsh restoration and treated wastewater discharge. In support of those activities, I am interested in a range of new sandbox features.
-Tides! Set one edge of the box to create a tidal boundary for the flow simulation, need to control amplitude and period.
– Companion water quality transport simulation. Color the water by a water quality constituent. A conservative tracer is a fine start. Will need to set the concentration for boundary flows. At least one for tidal boundary (salt water) and one for rain (fresh water).
– Point source and sink. It would be great to place a shape on the sand to create a point source or sink, like an arrow shape that creates a flow. Would need to control the flow rate and concentration for the point source. Right now I use a piece of Styrofoam on a stick to act as a rain cloud, but a shape on the sand surface would be better for a point source.
And a few more lower priority ideas.
– Monitoring station. Place a shape or some how identify a location in the sandbox that would then place a time series plot of water surface elevation or concentration on a separate monitor.
– Companion Particle Tracking Model. Allow sources of particles that would move with flow (or roll downhill, or otherwise wander about).
– Flow arrow or streak lines displayed on the flow animation.
– Color water by velocity.
I know these ideas are not easy to pull off. We have software developers in-house that may be able to help with a bit of direction.
Again, thanks for all your work on the ARSandbox, everyone that has seen ours has really enjoyed it.
John DeGeorgeMay 20, 2016 at 1:21 am #102201
Another vote for exporting the model in a STL format or a point cloud file, so we can manage to open in a 3D editing program. When I describe the AR Sandbox to others, the first question asked is if we can export for editing/3d printing/etc.August 23, 2016 at 4:34 pm #102307
I’d also like the ability to use the set up to record data and, as others have requested, export as a point cloud for post processing.October 17, 2016 at 3:08 am #102396
Thank you for developing and posting this amazing technology. It is incredible what technology is becoming accessible now at school level.
We are currently building our box, but can already see some things we would love to be able to do.
1) Wind and water current direction, so you can see the affect of changing wind direction on the direction and flow of water and waves.(yes this would also tie in with any erosion feature if you ever get that one going…that would be awesome!)
2) The ability to change runoff rates for different surface types. For example to allocate an area on the box as ‘forest’, or grassland, or bare ground, or rock and be able to show the differences in water run-off or ‘soak-in’ rates for the different ground covers.
3) Tides. For a coastal scenario. The ability to have a simulation of water level increase over time maybe as ‘simple’ as having a fluctuation altering the lower couple of colour graduations. Sorry I know nothing about programming unfortunately.
Thanks very much. I hope the value of what you are doing here translates into more funding for your project.October 17, 2016 at 5:41 am #102398
Just a disclaimer, I’m not a scientist or a researcher or anybody who would know anything about how this thing works… what I am is somebody who stumbled upon a video of this thing and immediately thought of a possible real-world application that could possibly justify more research funding…
See, I work in the construction industry and am currently studying construction management in college, and as I am currently learning, both in school and at work, to use different measurement devices and techniques, the video I saw of this device was like a eureka moment. The job that my company is just finishing up is a parking lot grading job, where it has been my job to use a rotary laser level to assist an excavator operator in regrading the sub-base elevation in preparation for interlocking stone. Although it is true that laser levels are a relatively new development in measurement equipment and a huge improvement over pre-existing solutions, the whole process is still rather cumbersome, owing to the fact that the elevation of the detector on the staff must be constantly readjusted based on the information provided by the drawings, and it doesn’t take much to become overwhelmed.
When I saw the video of your device, I immediately saw how incredibly useful this technology could be if it could be fashioned into a reliable, durable and easily set up system. Imagine a system that is a sort of combination between your device and a total station, which is a modern digital surveying device. Using GPS and other methods of determining position, elevation and angle, and perhaps a very bright projector or other method of relaying the necessary information, such a system could conceivably use information from a site drawing, which was likely created digitally in the first place, and display necessary changes in elevation RIGHT ON THE SOIL in real-time. Such a device would make the whole process of regrading very simple and intuitive, because it’s all visual rather than numerical, which would significantly reduce the amount of time required to complete the job, and trust me, anything that reduces the amount of time and work involved in completing a job in the construction industry is something that gets taken seriously and will find people willing to invest in.
I’m certainly not in a position to persue the development of such a device, I only offer my perspective as a possible source of funding for additional research. I can’t imagine a better use for the technology that you’ve created.
Thanks for reading and all the best.December 19, 2016 at 3:27 pm #102844
This is my first post on this forum.
I have done my first prototype and now i build my a nice second sandbox.
my wish is that i can split the system into a server/client system.
A client which is connected to the kinect and the projector.
A server with an powerful GPU to do the work.
I have tried to forward the USB subsiystem via usbip to my server and rewind the display via vnc.
forward the kinect didn’t work.
display via vnc works well.
The kinect sdk seems to have an network option.December 20, 2016 at 12:11 am #102845
The Kinect package has a client-server network protocol to stream 3D video data. SARndbox-2.3 does not have a client for network 3D cameras, but I just put it in for SARndbox-2.4.
The procedure will be to run a KinectServer daemon on the computer with the Kinect attached, and connect to that daemon from the computer running the SARndbox application.December 20, 2016 at 12:32 am #102846
that sounds perfect.
I will test it with an rasPi and a big server with a good GPU 🙂February 26, 2017 at 6:15 pm #103037
can add the function of the 3D print.For example, export the terrain to a file which can import to blender or other open source 3D software.And also add the main and important function which is to open the funds interface we can support this project to go ahead.
Thanks very much!
You must be logged in to reply to this topic.