Started in 2013, the Predictive Ecosystem Analyzer (PEcAn) effort aims to enhance ecosystem science, policy, and management by informing them with the best data and models available. An essential part of this vision is the synthesis of existing data sources with ecosystem models. As ecosystem science has seen an explosion of available data and data types there is a need to be able to ingest and adapt data from a wide variety of sources. In collaboration with the Brown Dog effort the PEcAn and Brown Dog teams work to support this data ingestion, helping scientists keep pace with the rate at which the research community is generating new observations. One place PEcAn uses Brown Dog is to handle the processing of input data used to drive models, such as meteorological observations, from a raw source to the format that a particular ecosystem model needs. Because Brown Dog is a cloud service, it not only does this transformation, but does so without using the computational or storage capacity of the user’s machine, making data more accessible to anyone regardless of the computational capabilities of their machine. This is critical for PEcAn users, who often only need a small local portion of a large (TB or more) global data set, which allows PEcAn to be run on laptops from field, where both storage and bandwidth are limited, and in the classroom. The ability of Brown Dog to record data provenance is also key for PEcAn users to ensuring data can be tracked down, scrutinized, and reused whenever needed. The PEcAn team will continue to build tools around Brown Dog in order to synthesize more environmental data with all ecosystem models and in turn make these data ingestion tools available to the larger ecological community through Brown Dog towards promoting reproducible ecosystem modeling and forecasting.
The Brown Dog project funds roughly a dozen Ph.D. and Masters students across its use cases to help drive activities within the supported scientific domains. Today we would like to congratulate Ankit Rai, Ph.D. who successfully defended his Ph.D. thesis this month. Ankit, working with Co-PI Barbara Minsker, has worked at the intersection of Informatics, Civil Engineering, and Social Science:
“My research work primarily addressed the limitation of current approach in studying landscape preferences by using advanced data science techniques. As a part of this work, a novel framework is created for identifying urban green storm water infrastructure (GI) designs (wetlands/ponds, urban trees, and rain gardens/bioswales) from high-resolution Google Earth images using state of the art computer vision and machine learning methods. The GI identification framework was also validated as an approach for collecting landscape preference data towards improving the understanding of what specific features are most desired. Previous research has shown that high-preference green settings are correlated with improved human health and well being. We further curated social media data using Twitter, Flickr, and Instagram to analyze GI preferences using qualitative codebook analysis and natural language processing techniques. The models and findings are implemented as Brown Dog services allowing others to leverage these tools as opposed to having to re-implement these capabilities within their research when using similar datasets”
As part of his research Ankit has developed a number of extractors to assign a green index to pedestrian routes based given path coordinates, automatically estimate human preference of landscapes given either images or text describing those landscapes, detect green infrastructure types within aerial images, as well as versions of these extractors capable of operating on data contained within social media feeds such as Twitter, Flickr, and Instagram.
To use these tools and more simply sign up for a Brown Dog account!:
Updated tutorial materials for the Brown Dog beta instance are now available here:
To facilitate users in the portions involving the creation of new transformation tools to be deployed in the Brown Dog Data Transformation Service (DTS) we have incorporated a new docker based utility that allows one to get going quickly, the BD Development Base:
On your local machine all you need is git and docker (which you would need regardless), then simply type:
and you will be up and running with your own DTS instance (and all of its dependent sub-services). To startup a basic tool development environment automatically connected to this personal DTS instance then type:
We also provide a Virtual Machine containing these components and utilities to further facilitate users as they get started:
If interested in an onsite hands-on tutorial for your organization please let us know at firstname.lastname@example.org!
Created a new Brown Dog YouTube video highlighting a number of the client interfaces.
We had a paper accepted into IEEE Big Data this year. The paper goes over the architecture as well as the various components that make up Brown Dog. If you need to cite Brown Dog this is the paper to use:
S. Padhy, G. Jansen, J. Alameda, E. Black, L. Diesendruck, M. Dietze, P. Kumar, R. Kooper, J. Lee, R. Liu, R. Marciano, L. Marini, D. Mattson, B. Minsker, C. Navarro, M. Slavenas, W. Sullivan, J. Votava, K. McHenry, "Brown Dog: Leveraging Everything Towards Autocuration", IEEE Big Data, 2015
Moving into a friendly user mode we have begun to put together some new information resources to aid new users in using Brown Dog services. For example the Brown Dog flyer below also serves as a cheat containing the REST API for conversions/extractions and details on how to add data analysis/manipulation tools as new converters/extractors to extend Brown Dog's capabilities.
Our ultimate goal will be to include support for the DAP and DTS at the OS level and like the DNS, though largely invisible to the average user, be an essential part of the internet in terms of accessing, finding, and using data.
Here is a video of some of the capabilities we are adding to Brown Dog in support of our Critical Zone use case with Praveen Kumar at UIUC. Towards better understanding the human impacts to a regions hydrology we have added extraction tools to identify floodplains in lidar data and pull historical river from digitized maps:
Special thanks to Qina Yan (Ph.D. student in Civil & Environmental Engineering), Liana Diesendruk (Research Programmer), and Smruti Padhy (Research Programmer), Jong Lee (Senior Research Programmer), and Chris Navarro (Senior Research Programmer), for their work on these tools and for developing the DataWolf workflow.
Coming releases of PEcAn will come with the ability to utilize Brown Dog for data conversions to model specific formats (part of our Ecology use case with Mike Dietze at Boston University). In developments to come this will aid PEcAn in the usage of large data sets such as NARR. Below you can see an example of Brown Dog being used to process the data needed for the SIPNET model:
and for the ED model:
Special thanks to Betsy Cowdery (Ph.D. student in Earth Sciences), Edgar Black (Research Programmer), and Rob Kooper (Senior Research Programmer) for adding the needed conversions to Brown Dog and interfacing to PEcAn.
Added some some videos showing what we are doing for our Green Infrastructure use case with Barbara Minsker, Art Schmidt, and Bill Sullivan at UIUC. In an effort to study the human health impacts of our day to day environment a number of extraction tools were built and deployed to do things such as assign a green index to photographed scenes, estimate human preference, estimate human sentiment from written descriptions, and identify bodies of water in aerial photos. Below we show the extraction of green indices for various urban pathways:
the extraction of a human preference score estimate for photographed scenes:
the extraction of the writers sentiment for text descriptions of a given scene:
and for the extraction of areas containing bodies of water in aerial photos:
Special thanks to Ankit Rai (Ph.D. student in Informatics), Marcus Slavenas (Research Programmer), and Luigi Marini (Senior Research Programmer) for their development of these extractors and example workflows in Clowder.
In the next video we show how to use the extracted data to build an index over a collection of files and then perform a search to find similar data:
The Brown Dog Tools Catalog will serve as a means to both collect new conversion/extraction tools from the community while simultaneously serving as platform for finding and preserving such tools. We have populated the Tools Catalog with a handful of the Medici extractors and Polyglot conversion scripts to get it started:
Adding a new tool is a simple matter of registering the tool along with a very simple wrapper script that will allow either the DAP or DTS to make use of it. Below we show an example adding scripts to make use of PEcAn for ecological model conversions:
and converting that to the format needed by various models, e.g. SIPNET:
The DAP, built on top of NCSA Polyglot, chains conversion tools together within the "cloud" allowing one to jump from format to format as needed. The DAP will eventually support the intelligent moving of data/computation to handle large datasets (e.g. NARR) and support a variety of models within ecology (e.g. ED) as well as other domains.
Some videos of the new Brown Dog Google Chrome extension allowing the Data Tilling Service (DTS), based off of NCSA Medici, to be called on arbitrary pages in order to index collections of data. Note the text in the queries is not part of the page or images on the page but being extracted from the image contents using cloud hosted tools within the DTS, specifically the face detector within OpenCV and the Tesserct OCR engine. The DTS will host a suite of such tools and make it easy for users to add additional tools:
Added some higher resolution videos of the DAP bookmarklet being used for images:
for 3D data:
and for archive/container files (e.g. zip, rar):