Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions site/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,10 @@ <h2>Standard</h2>
projects, including Calcite, Cassandra, Drill, Hadoop, HBase, Ibis,
Impala, Kudu, Pandas, Parquet, Phoenix, Spark, and Storm making it
the de-facto standard for columnar in-memory analytics.</p>

<p>Learn more about projects that are <a href="{{ site.baseurl
}}/powered_by/">Powered By Apache Arrow</a></p>

</div>
</div> <!-- close "row" div -->

Expand Down
45 changes: 18 additions & 27 deletions site/powered_by.md
Original file line number Diff line number Diff line change
Expand Up @@ -43,11 +43,9 @@ names, etc.) like "arrow-foo". These are permitted. Nominative use of trademarks
in descriptions is also always allowed, as in "BigCoProduct is a widget for
Apache Arrow".

### Open Source Projects

To add yourself to the list, please email [email protected] with your
To add yourself to the list, please open a pull request adding your
organization name, URL, a list of which Arrow components you are using, and a
short description of your use case.
short description of your use case. See the following for some examples.

* **[Apache Parquet][3]:** A columnar storage format available to any project
in the Hadoop ecosystem, regardless of the choice of data processing
Expand All @@ -61,10 +59,23 @@ short description of your use case.
* **[Dask][15]:** Python library for parallel and distributed execution of
dynamic task graphs. Dask supports using pyarrow for accessing Parquet
files
* **[Dremio][9]:** A self-service data platform. Dremio makes it easy for
users to discover, curate, accelerate, and share data from any source.
It includes a distributed SQL execution engine based on Apache Arrow.
Dremio reads data from any source (RDBMS, HDFS, S3, NoSQL) into Arrow
buffers, and provides fast SQL access via ODBC, JDBC, and REST for BI,
Python, R, and more (all backed by Apache Arrow).
* **[GeoMesa][8]:** A suite of tools that enables large-scale geospatial query
and analytics on distributed computing systems. GeoMesa supports query
results in the Arrow IPC format, which can then be used for in-browser
visualizations and/or further analytics.
* **[GOAI][19]:** Open GPU-Accelerated Analytics Initiative for Arrow-powered
analytics across GPU tools and vendors
* **[Graphistry][18]:** Supercharged Visual Investigation Platform used by
teams for security, anti-fraud, and related investigations. The Graphistry
team uses Arrow in its NodeJS GPU backend and client libraries, and is an
early contributing member to GOAI and Arrow\[JS\] focused on bringing these
technologies to the enterprise.
* **[libgdf][14]:** A C library of CUDA-based analytics functions and GPU IPC
support for structured data. Uses the Arrow IPC format and targets the Arrow
memory layout in its analytic functions. This work is part of the [GPU Open
Expand All @@ -75,6 +86,9 @@ short description of your use case.
* **[pandas][12]:** data analysis toolkit for Python programmers. pandas
supports reading and writing Parquet files using pyarrow. Several pandas
core developers are also contributors to Apache Arrow.
* **[Quilt Data][13]:** Quilt is a data package manager, designed to make
managing data as easy as managing code. It supports Parquet format via
pyarrow for data access.
* **[Ray][5]:** A flexible, high-performance distributed execution framework
with a focus on machine learning and AI applications. Uses Arrow to
efficiently store Python data structures containing large arrays of numerical
Expand All @@ -91,29 +105,6 @@ short description of your use case.
Arrow Tables and RecordBatches in addition to the Python Database API
Specification 2.0.

### Companies and Organizations

To add yourself to the list, please email [email protected] with your
organization name, URL, a list of which Arrow components you are using, and a
short description of your use case.

* **[Dremio][9]:** A self-service data platform. Dremio makes it easy for
users to discover, curate, accelerate, and share data from any source.
It includes a distributed SQL execution engine based on Apache Arrow.
Dremio reads data from any source (RDBMS, HDFS, S3, NoSQL) into Arrow
buffers, and provides fast SQL access via ODBC, JDBC, and REST for BI,
Python, R, and more (all backed by Apache Arrow).
* **[GOAI][19]:** Open GPU-Accelerated Analytics Initiative for Arrow-powered
analytics across GPU tools and vendors
* **[Graphistry][18]:** Supercharged Visual Investigation Platform used by
teams for security, anti-fraud, and related investigations. The Graphistry
team uses Arrow in its NodeJS GPU backend and client libraries, and is an
early contributing member to GOAI and Arrow\[JS\] focused on bringing these
technologies to the enterprise.
* **[Quilt Data][13]:** Quilt is a data package manager, designed to make
managing data as easy as managing code. It supports Parquet format via
pyarrow for data access.

[1]: https://www.apache.org/foundation/marks/
[2]: https://www.apache.org/foundation/marks/faq/
[3]: https://parquet.apache.org/
Expand Down