# elasticsearch-VR **Repository Path**: mirrors_mozilla/elasticsearch-VR ## Basic Information - **Project Name**: elasticsearch-VR - **Description**: INACTIVE - http://mzl.la/ghe-archive - experimental VR interface to elastic search - **Primary Language**: Unknown - **License**: MPL-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2020-08-22 - **Last Updated**: 2026-03-29 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # elasticsearch-VR What ----- Experimental VR interface to elastic search. When ---- Project for the Mozilla winter break of 2017. Why ---- VR scenes are fun to interact with, even more fun when they do real things like interface to other systems. This is an experiment to see if VR can be a useful interface into Elastic Search for discovery, querying, health/status, etc. Stack ------ - a-frame.io for the webVR portion - Windows Mixed Reality Headset - Elastic search (in docker) for a target ES - Python/flask for the middle/data driving tier. The idea is that you would use the Mixed Reality headset to enter the scene and trigger basic interactions like: - Connect to a cluster - Visualize the cluster size/health - Select indexes - Get a visual representation of the index documents/fields - Use the data dictionary for the index to craft queries - Visualize the results as aggregations, details, etc. What works ----------- Basic VR scene loading. QuickStart ----------- Set the environment ``` set -a;source env.example;set +a ``` Start the app ``` python app.py ``` Use a WebVR enabled browser (firefox/IE) to navigate to: ``` http://localhost:5000/ ``` Open the VR scene using whatever you have handy for VR, ideally one of - Windows Mixed Reality Controller (since that's what it's tested with)