Connecting to a Matillion ETL Instance (EC2)

From a web browser

These instructions assume you have already successfully launched Matillion ETL from the AWS marketplace. See here for detail.

Once the Amazon EC2 Instance is started all you should need is a web browser to connect to Instance, and then you can get started building ETL jobs.

These instructions assume you have already successfully launched Matillion ETL from the AWS marketplace. See here for detail.


Once the Amazon EC2 Instance is started all you should need is a web browser to connect to Instance, and then you can get started building ETL jobs.

These instructions assume you have already successfully launched Matillion ETL from the GCP marketplace. See here for detail.

 

Once the Amazon EC2 Instance is started all you should need is a web browser to connect to Instance, and then you can get started building ETL jobs.

  • The URL to connect to is: http://<Host Name or IP address>/
  • The host name or IP address can be found in the EC2 Management Console in Services -> EC2 -> Instances
  • You should then see a login screen such as.  



Note: When starting a new instance the default username is ec2-user and the default password is the instance ID i-xxxxxxxx (e.g. i-88ed92c6)

Alternatively, if the instance is public then use the public DNS as below.

Initial Environment Set-up

To get started building ETL Jobs in Matillion ETL, the final step is to get connected to your Redshift cluster. If you attached an IAM Role to your AMI, Matillion ETL can interrogate your cluster to discover the basic connection information, otherwise it will need to be set manually.

To get started building ETL Jobs in Matillion ETL, the final step is to get connected to your Snowflake account.

To get started building ETL Jobs in Matillion ETL, the final step is to get connected to your BigQuery cluster.


On first run, the product will take you to the Create Project screen:
 
  • Project Group: An existing or new Group (of arbitrary name) used for logically grouping Projects.
  • Project Name: An arbitrary name for your new Project
  • Project Description: A searchable description of your project.
  • Include Samples: Deselect if you do not want to include sample jobs in the project.

Environment details requires information about the RedshiftSnowflakeBigQuery account.



  • Name: Describes the environment. e.g. Test/Live etc
  • Credentials: Choose a set of credentials to use with this Project.
  • Cluster: If you have created an IAM Role, this box will list your Redshift clusters that are visible to that role in the same region as the Instance is running. Selecting that cluster will populate the remaining fields on the screen with sensible defaults. If the box is empty, these will need to be set manually.


Connection Settings:

  • Default Schema: 'Public' by default, but if you have configured multiple schemas within your Redshift database, specify the schema you would like to use here.
  • Endpoint: The physical address of the Leader Node. It will be either a name or an I.P. address.
  • Port:  Usually 5439 or 5432, but could be configured differently when setting up your Redshift cluster.
  • Database:  However you named it when setting up your Redshift cluster. You may run with multiple database names; in which case, choose the one you would like to use for this project here.
  • Username: Username for the environment connection.
  • Password: Password for the environment connection.
  • Enable SSL: To encrypt communications between Matillion and Redshift. Some Redshift clusters may be configured to require this.
  • Account: Your Snowflake account name. This is part of the URL you use to login to Snowflake. e.g. <AccountName>.snowflakecomputing.com
  • Default Warehouse: The name of the Warehouse on your Snowflake account to use by default.
  • Default Database: The name of the Database on your Snowflake account to use by default.
  • Default Schema:  'public' by default, but if you have configured multiple schemas within your Redshift database, specify the schema you would like to use here.
  • Default Role: The default Snowflake Role that this Project assumes.
  • Username: Username for the environment connection.
  • Password: Password for the environment connection.
  • Enable SSL: To encrypt communications between Matillion and Redshift. Some Redshift clusters may be configured to require this.
 
  • Default Project: The name of the Project on your GCP account to use by default.
  • Default Dataset: The name of the Dataset on your GCP account to use by default.

Click Test and ensure connection is successful.

Finally, click OK to create your project and environment. Matillion ETL will launch the sample job. To get started building your own job, right click on the Explorer panel. 
 

To get started building ETL Jobs in Matillion ETL the final step is to get connected to your Snowflake account. To do so, please refer to the documentation on Connecting To Snowflake.