A new customer table is created by a data pipeline in a Snowflake schema where MANAGED ACCESS enabled.
.... Can gran access to the CUSTOMER table? (Select THREE.)
The roles that can grant access to the CUSTOMER table are the role that owns the schema, the role that owns the database, and the SECURITYADMIN role. These roles have the ownership or the manage grants privilege on the schema or the database level, which allows them to grant access to any object within them. The other options are incorrect because they do not have the necessary privilege to grant access to the CUSTOMER table. Option C is incorrect because the role that owns the customer table cannot grant access to itself or to other roles. Option D is incorrect because the SYSADMIN role does not have the manage grants privilege by default and cannot grant access to objects that it does not own. Option F is incorrect because the USERADMIN role with the manage grants privilege can only grant access to users and roles, not to tables.
Which methods will trigger an action that will evaluate a DataFrame? (Select TWO)
The methods that will trigger an action that will evaluate a DataFrame are DataFrame.collect() and DataFrame.show(). These methods will force the execution of any pending transformations on the DataFrame and return or display the results. The other options are not methods that will evaluate a DataFrame. Option A, DataFrame.random_split(), is a method that will split a DataFrame into two or more DataFrames based on random weights. Option C, DataFrame.select(), is a method that will project a set of expressions on a DataFrame and return a new DataFrame. Option D, DataFrame.col(), is a method that will return a Column object based on a column name in a DataFrame.
What is a characteristic of the use of external tokenization?
External tokenization is a feature in Snowflake that allows users to replace sensitive data values with tokens that are generated and managed by an external service. External tokenization allows the preservation of analytical values after de-identification, such as preserving the format, length, or range of the original values. This way, users can perform analytics on the tokenized data without compromising the security or privacy of the sensitive data.
A CSV file around 1 TB in size is generated daily on an on-premise server A corresponding table. Internal stage, and file format have already been created in Snowflake to facilitate the data loading process
How can the process of bringing the CSV file into Snowflake be automated using the LEAST amount of operational overhead?
This option is the best way to automate the process of bringing the CSV file into Snowflake with the least amount of operational overhead. SnowSQL is a command-line tool that can be used to execute SQL statements and scripts on Snowflake. By scheduling a SQL file that executes a PUT command, the CSV file can be pushed from the on-premise server to the internal stage in Snowflake. Then, by creating a pipe that runs a COPY INTO statement that references the internal stage, Snowpipe can automatically load the file from the internal stage into the table when it detects a new file in the stage. This way, there is no need to manually start or monitor a virtual warehouse or task.
A Data Engineer defines the following masking policy:
....
must be applied to the full_name column in the customer table:
Which query will apply the masking policy on the full_name column?
The query that will apply the masking policy on the full_name column is ALTER TABLE customer MODIFY COLUMN full_name SET MASKING POLICY name_policy;. This query will modify the full_name column and associate it with the name_policy masking policy, which will mask the first and last names of the customers with asterisks. The other options are incorrect because they do not follow the correct syntax for applying a masking policy on a column. Option B is incorrect because it uses ADD instead of SET, which is not a valid keyword for modifying a column. Option C is incorrect because it tries to apply the masking policy on two columns, first_name and last_name, which are not part of the table structure. Option D is incorrect because it uses commas instead of dots to separate the database, schema, and table names
Mohammad
11 minutes agoCarmelina
9 days agoBelen
23 days agoAnnabelle
28 days agoWilbert
2 months agoLettie
3 months agoAlonso
3 months agoViola
4 months agoKayleigh
4 months agoRozella
4 months agoHana
5 months agoRasheeda
5 months agoLashandra
5 months agoTalia
6 months agoJarod
6 months agoMarion
6 months agoGilberto
6 months agoZack
7 months agoIvory
7 months agoJustine
7 months agoCarey
7 months agoChantay
7 months agoGerald
7 months agoAsha
8 months agoLucia
8 months agoClaribel
8 months agoJohnathon
8 months agoEvette
8 months agoLavelle
9 months agoCarin
9 months agoAretha
9 months agoWilliam
9 months agoAnnita
10 months agoYolando
10 months agoReita
10 months agoSalena
10 months agoCeola
10 months agoLonny
12 months agoGerri
1 years agoRolland
1 years agoJolene
1 years agoFatima
1 years agoPa
1 years ago