You are a data scientist using Oracle AutoML to produce a model and you are evaluating the score metric for the model. Which of the following TWO prevailing metrics would you use for evaluating multiclass classification model?
A. Recall
B. Mean squared error
C. F1 Score
D. R-Squared
E. Explained variance score
You are working as a data scientist for a healthcare company. They decide to analyze the data to find patterns in a large volume of electronic medical records. You are asked to build a PySpark solution to analyze these records in a JupyterLab notebook. What is the order of recommended steps to develop a PySpark application in Oracle Cloud Infrastructure (OCI) Data Science?
A. Launch a notebook session. Configure core-site.xml. Install a PySPark conda environ- ment. B. Develop your PySpark application Create a Data Flow application with the Ac- celerated Data Science (ADS) SOK
B. Configure core-site.xml. Install a PySPark conda environment. Create a Data Flow application with the Accelerated Data Science (ADS) SDK Develop your PySpark ap- plication. Launch a notebook session.
C. Launch a notebook session. Install a PySpark conda environment. Configure coresite.xml.
D. Develop your PySpark application. Create a Data Flow application with the Ac-celerated Data science (ADS) SDK.
E. Install a spark conda environment. Configure core-site.xml. Launch a notebook session:Create a Data Flow application with the Accelerated Data Science (ADS) SOK. Develop your PySpark application
As a data scientist, you are working on a global health data set that has data from more than 50 countries. You want to encode three features, such as 'countries', 'race', and 'body organ' as categories. Which option would you use to encode the categorical feature?
A. DataFramLabelEncode()
B. auto_transform()
C. OneHotEncoder()
D. show_in_notebook()
You realize that your model deployment is about to reach its utilization limit. What would you do to avoid the issue before requests start to fail?
A. Update the deployment to use fewer instances.
B. Delete the deployment.
C. Reduce the load balancer bandwidth limit so that fewer requests come in.
D. Update the deployment to use a larger virtual machine (mare CPUs/memory).
E. Update the deployment to add more instances.
You have just received a new data set from a colleague. You want to quickly find out summary information about the data set, such as the types of features, total number of observations, and datadistributions, Which Accelerated Data Science (ADS) SDK method from the ADandDataset class would you use?
A. Show_in_notebook{}
B. To_xgb{}
C. Compute{}
D. Show_corr{}
You are asked to prepare data for a custom-built model that requires transcribing Spanish video recordings into a readable text format with profane words identified. Which Oracle Cloud service would you use?
A. OCI Translation
B. OCI Language
C. OCI Anomaly Detection
D. OCI Speech
As you are working in your notebook session, you find that your notebook session does not have enough compute CPU and memory for your workload. How would you scale up your notebook session without losing your work?
A. Ensure your files and environments are written to the block volume storage under the /home/datascience directory, deactivate the notebook session, and activate the notebook larger compute shape selected.
B. Down your files and data to your local machine, delete your notebook session, provision tebook session on a larger compute shape, and upload your files from your local the new notebook session.
C. Deactivate your notebook session, provision a new notebook session on larger compute shape, and re-create all your file changes.
D. Create a temporary bucket in Object Storage, write all your files and data to Object Storage, delete tur ctebook session, provision a new notebook session on a larger com-pute shape, and capy your flies and data from your temporary bucket onto your new notebook session.
For your next data science project, you need access to public geospatial images. Which Oracle Cloud service provides free access to those images?
A. Oracle Big Data Service
B. Oracle Analytics Claud
C. Oracle Cloud Infrastructure (OCI) Data Science
D. Oracle Open Data
You want to ensure that all stdout and stderr from your code are automatically collected and logged, without implementing additional logging in your code. How would you achieve this with Data Science Jobs?
A. Data Science Jots does not support automatic fog collection and storing.
B. On job creation, enable logging and select a log group. Then, select either log or the op- tion to enable automatic log creation.
C. You can implement custom logging in your code by using the Data Science Jobs logging.
D. Make sure that your code is using the standard logging library and then store all the logs to Check Storage at the end of the job.
You have a complex Python code project that could benefit from using Data Science Jobs as it is a repeatable machine learning model training task. The project contains many subfolder and classes. What is the best way to run this project as a job?
A. ZIP the entire code project folder, upload it as a Job artifact on job creation and set JOB_RUN_ENTRYPOINT to point to the main executable file.
B. ZIP the entire code project folder and upload it as a Job artifact on job creation, Jobs identities the main executable file automatically.
C. Rewrite your code so that a single executable Python or Bash/Shell script file.
D. ZIP the entire code project folder and upload it as a Job artifact Jobs automatically identifies That main top level where the code is run.