Presto - Superset
Presto | Distributed SQL Query Engine for Big Data
Distributed SQL Query Engine for Big Data
prestosql.io
https://chanchanhome.tistory.com/44
검색엔진 프레스토 Presto 알아보기
검색엔진으로 Redshift를 사용하려고 했는데 왜 이렇게 속도가 안나오는거지? Redshift는 속도 보다는 대용량 데이터를 보관하고 찾을 수 있는 DW로 만족해야하는 것 같다. ㅠㅜ 데이터를 빠르고 여러가지 방법으..
chanchanhome.tistory.com
https://medium.com/@ian.nam.kr/apache-presto-benchmark-b726b6283e0e
[BigData] Apache Presto Benchmark
About Presto
medium.com
https://support.treasuredata.com/hc/en-us/articles/360000708727-Presto-JDBC-Connection
Presto JDBC Connection
Connect to Arm Treasure Data’s Presto using our JDBC Driver. Setup Download presto-jdbc driver For using secure HTTPS (port 443) connection you need to use presto 0.148 or later. Use the follo...
support.treasuredata.com
https://hub.docker.com/r/starburstdata/presto/
Docker Hub
hub.docker.com
https://github.com/1900/presto-spring-boot-starter
1900/presto-spring-boot-starter
Spring Boot Starter for Presto. Contribute to 1900/presto-spring-boot-starter development by creating an account on GitHub.
github.com
https://inventa.live/artifactory/gradle-release/com/facebook/presto/presto-kudu/
Index of gradle-release/com/facebook/presto/presto-kudu
inventa.live
CREATE TABLE orders ( orderkey bigint, custkey bigint, orderstatus varchar(1), totalprice double, orderdate date, orderpriority varchar(15), clerk varchar(15), shippriority integer, comment varchar(79) ) WITH ( external_location = 's3://my_bucket/path/to/folder', format = 'ORC' -- or 'PARQUET' );
https://stackoverflow.com/questions/53392430/correct-way-to-restart-presto-server-service-on-emr
Correct way to restart presto-server service on EMR
I have to restart presto-server on EMR to load my plugin. With reference to official AWS EMR docs: On some places they say sudo restart presto-server On others they say, sudo stop presto-...
stackoverflow.com
https://stackoverflow.com/questions/27867158/presto-query-no-worker-nodes-available
Presto - Query ... No worker nodes available
Using Amazon EMR, Hive .13, Hadoop 2.x, and Presto Server 0.89. Trying to set up Presto to query data that is usually queried through Hive. Hive metadata is stored in MySQL. Presto Server is inst...
stackoverflow.com
insert into hive.default.tbl_adobe_data (visits) select visits from mysql.insight.tbl_adobe_data;
https://blog.wisen.co.kr/?p=11295
AWS EMR Series - Apache Superset으로 presto 쿼리 결과 시각화
Cloud Expert Group - GS네오텍
blog.wisen.co.kr
Presto, Zeppelin을 이용한 초간단 BI 시스템 구축 사례(2) | Popit
Presto, Zeppelin을 이용한 초간단 BI 시스템 구축 사례
www.popit.kr
https://docs.aws.amazon.com/ko_kr/emr/latest/ReleaseGuide/presto-adding-db-connectors.html
데이터베이스 커넥터 추가 - Amazon EMR
PostgreSQL 커넥터가 설치되고 구성된 클러스터를 시작하려면 먼저 다음 콘텐츠가 있는 구성 분류(예: myConfig.json)를 지정하는 JSON 파일을 생성하여 로컬에 저장합니다. Presto 설명서의 PostgreSQL 커넥터 항목에 표시된 대로 연결 속성을 설정에 맞게 바꿉니다. [ { "Classification": "presto-connector-postgresql", "Properties": { "connection-url": "jdbc
docs.aws.amazon.com
https://stackoverflow.com/questions/53392430/correct-way-to-restart-presto-server-service-on-emr
Correct way to restart presto-server service on EMR
I have to restart presto-server on EMR to load my plugin. With reference to official AWS EMR docs: On some places they say sudo restart presto-server On others they say, sudo stop presto-...
stackoverflow.com
https://github.com/prestodb/presto/issues/7432
"No nodes available to run query" on EMR using presto-mysql-connector · Issue #7432 · prestodb/presto
Hi all, I'm setting up EMR with 1 Master node, 2 Core nodes. When using presto-mysql-connector plugin, i queried data on presto which connect to the table on Mysql and it is fail. select * from...
github.com
* Mysql 설정
sudo vi /etc/presto/conf/catalog/mysql.properties
connector.name=mysql
connection-url=jdbc:mysql://dataview.cangewyjzrdn.ap-northeast-1.rds.amazonaws.com:3306
connection-user=mm
connection-password=442423!!f
* 설정은 모든 노드(마스터+슬레이브) 포함
ssh -i ~/.ssh/zeliter.pem hadoop@ec2-18-182-48-214.ap-northeast-1.compute.amazonaws.com
ssh -i ~/.ssh/zeliter.pem hadoop@ec2-52-194-239-82.ap-northeast-1.compute.amazonaws.com
ssh -i ~/.ssh/zeliter.pem hadoop@ec2-13-231-141-26.ap-northeast-1.compute.amazonaws.com
sudo restart presto-server
presto-cli
* 설정 후 확인
select * from mysql.zeliter.review limit 10;
CREATE EXTERNAL TABLE member_review (
member_id STRING,
review_desc STRING
)
STORED AS PARQUET
LOCATION 's3://newzeliter2/review/parquet';
insert into hive.default.member_review (member_id, review_desc) select cast(member_id as varchar(100)), content from mysql.zeliter.review;
select * from hive.default.member_review limit 10;
* CSV 및 S3 를 통한 테이블 생성
Then you can such table in Presto Hive connector with:
CREATE TABLE hive.default.csv_table_with_custom_parameters ( c_bigint varchar, c_varchar varchar) WITH ( csv_escape = '', csv_quote = '', csv_separator = U&'\0001', -- to pass unicode character external_location = 'hdfs://hadoop/datacsv_table_with_custom_parameters', format = 'CSV')
In your case it would be:
CREATE TABLE hive.default.csv_table_with_custom_parameters ( PassengerId int, Survived int, Pclass int, Name string, Sex string, Age int, SibSp int, Parch int, Ticket int, Fare double, Cabin string, Embarked string) WITH ( csv_escape = '\', csv_quote = '"', csv_separator = ',', external_location = 's3://my_bucket/titanic_train/', format = 'CSV')