Weekend Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmas50

Oracle 1z0-1110-25 - Oracle Cloud Infrastructure 2025 Data Science Professional

Page: 2 / 5
Total 158 questions

You are working as a data scientist for a healthcare company. They decided to analyze the data to find patterns in a large volume of electronic medical records. You are asked to build a PySpark solution to analyze these records in a JupyterLab notebook. What is the order of recommended steps to develop a PySpark application in OCI Data Science?

A.

Launch a notebook session, configure core-site.xml, install a PySpark conda environment, develop your PySpark application, create a Data Flow application with the Accelerated Data Science (ADS) SDK

B.

Configure core-site.xml, install a PySpark conda environment, create a Data Flow application with the Accelerated Data Science (ADS) SDK, develop your PySpark application, launch a notebook session

C.

Install a Spark conda environment, configure core-site.xml, launch a notebook session, create a Data Flow application with the Accelerated Data Science (ADS) SDK, develop your PySpark application

D.

Launch a notebook session, install a PySpark conda environment, configure core-site.xml, develop your PySpark application, create a Data Flow application with the Accelerated Data Science (ADS) SDK

You are a data scientist working inside a notebook session and you attempt to pip install a package from a public repository that is not included in your conda environment. After running this command, you get a network timeout error. What might be missing from your networking configuration?

A.

FastConnect to an on-premises network

B.

Primary Virtual Network Interface Card (VNIC)

C.

NAT Gateway with public internet access

D.

Service Gateway with private subnet access

You have an image classification model in the model catalog which is deployed as an HTTP endpoint using model deployments. Your tenancy administrator is seeing increased demands and has asked you to increase the load balancing bandwidth from the default of 10Mbps. You are provided with the following information:

    Payload size in KB = 1024

    Estimated requests per second = 120 requests/second (Monday through Friday, in every month, in every year)

    Buffer percentage = 20%What is the optimal load balancing bandwidth to redeploy your model?

A.

452 Mbps

B.

52 Mbps

C.

7052 Mbps

D.

1152 Mbps

Which stage in the machine learning life cycle helps in identifying the imbalance present in the data?

A.

Data Modeling

B.

Data Monitoring

C.

Data Exploration

D.

Data Access

Where do calls to stdout and stderr from score.py go in the model deployment?

A.

The file that was defined for them on the virtual machine (VM)

B.

The OCI Console

C.

The OCI Cloud Shell, which can be accessed from the console

D.

The predict log in the Oracle Cloud Infrastructure (OCI) Logging service as defined in the deployment

You realize that your model deployment is about to reach its utilization limit. What would you do to avoid the issue before requests start to fail? Pick THREE.

A.

Update the deployment to add more instances

B.

Delete the deployment

C.

Update the deployment to use fewer instances

D.

Update the deployment to use a larger virtual machine (more CPUs/memory)

E.

Reduce the load balancer bandwidth limit so that fewer requests come in

As a data scientist for a hardware company, you have been asked to predict the revenue demand for the upcoming quarter. You develop a time series forecasting model to analyze the data. Select the correct sequence of steps to predict the revenue demand values for the upcoming quarter.

A.

Verify, prepare model, deploy, save, predict

B.

Predict, deploy, save, verify, prepare model

C.

Prepare model, deploy, verify, save, predict

D.

Prepare model, verify, save, deploy, predict

On which option do you set Oracle Cloud Infrastructure Budget?

A.

Compartments

B.

Instances

C.

Free-form tags

D.

Tenancy

As a data scientist, you are tasked with creating a model training job that is expected to take different hyperparameter values on every run. What is the most efficient way to set those parameters with Oracle Data Science Jobs?

A.

Create a new job every time you need to run your code and pass the parameters as environment variables

B.

Create your code to expect different parameters as command-line arguments, and create a new job every time you run the code

C.

Create a new job by setting the required parameters in your code, and create a new job for every code change

D.

Create your code to expect different parameters either as environment variables or as command-line arguments, which are set on every job run with different values

Arrange the following in the correct Git Repository workflow order:

    Install, configure, and authenticate Git.

    Configure SSH keys for the Git repository.

    Create a local and remote Git repository.

    Commit files to the local Git repository.

    Push the commit to the remote Git repository.

A.

2, 3, 1, 4, 5

B.

4, 2, 3, 1, 5

C.

3, 5, 1, 2, 4

D.

1, 2, 3, 4, 5