The following diagram shows the overall architecture for Amundsen.
The frontend service serves as web UI portal for users interaction. It is Flask-based web app which representation layer is built with React with Redux, Bootstrap, Webpack, and Babel.
The search service proxy leverages Elasticsearch's search functionality (or Apache Atlas's search API, if that's the backend you picked) and provides a RESTful API to serve search requests from the frontend service. This API is documented and live explorable through OpenAPI aka "Swagger". Currently only table resources are indexed and searchable. The search index is built with the databuilder elasticsearch publisher.
The metadata service currently uses a Neo4j proxy to interact with Neo4j graph db and serves frontend service's metadata. The metadata is represented as a graph model: The above diagram shows how metadata is modeled in Amundsen.
Amundsen provides a data ingestion library for building the metadata. At Lyft, we build the metadata once a day using an Airflow DAG (examples).
In addition to "real use" the databuilder is also employed as a handy tool to ingest some "pre-cooked" demo data used in the Quickstart guide. This allows you to have a supersmall sample of data to explore so many of the features in Amundsen are lit up without you even having to setup any connections to databases etc. to ingest real data.