Suppose that a given dose-response experiment the dose of po…

Suppose that a given dose-response experiment the dose of poison X and proportion m of insects out of M that are killed at each dose are given. A logistic regression model was used to model probability of killing  insects [P(Y=1)]. Fitted logistic model is given by  l o g i t ( P ( Y = 1 ) ) = β 0 + β 1 X . {“version”:”1.1″,”math”:”logit(P(Y=1))=\beta_0+\beta_1 X.”} Compute the effective dose needed to kill 60% of the insect population.

Choose the correct answer: 1. Откуда Мария? A) из России B)…

Choose the correct answer: 1. Откуда Мария? A) из России B) из Англии C) из США D) из Германии 2. Где Мария изучает русский язык? A) В Лондоне B) В Москве C) В Санкт-Петербурге D) В Новосибирске 3. Где Мария завтракает обычно? A) На кухне B) На балконе C) В кафе D) В гостиной 4. Что Мария обычно ест на завтрак? A) Блины с медом B) Яичницу с беконом и помидорами, тост и фасоль C) Борщ со сметаной и хлебом D) Равиоли с сыром 5. Что Мария пьёт на завтрак? A) Кофе B) Чёрный чай C) Сок D) Минеральную воду 8. Как Мария и её сосед завтракают вместе? A) На одном балконе B) На разных балконах C) В кафе D) В парке9. На каком языке они разговаривают? A) Только по-русски B) Только по-английски C) Иногда по-русски, иногда по-английски D) На французском 10. Что сосед приготовил для Марии? A) Борщ B) Пельмени C) Блины со сметаной D) Кашу с яйцом 11. Что Мария сказала о русской еде? A) Сказала, что невкусно B) Очень вкусно! C) Не хочу пробовать D) Не любит русскую кухню

Download the starter notebook for this exam here: mid2026.ip…

Download the starter notebook for this exam here: mid2026.ipynb If your browser shows the file inline, you can right click on the displayed file and choose  “Save As” to save a copy of this ipynb file.  Important: please save it as a new name  mid2026-your-name.ipynb before uploading it to Databricks. Use the file upload box below to submit your answer (HTML export of your Databricks notebook) 2026 MSBA 6331 Midterm Exam, Part 2 (40 points) Name __________________________________________ Read these instructions first! Download mid2026.ipynb and RENAME it to mid2026-your-name.ipynb (replace “your-name” with your actual name) before uploading it to Databricks. Answer your questions using mid2026-your-name.ipynb. The notebook has no question text on purpose – we try to minimize the risk of exam leakage. You must enter your answers in the right cells as identified by sections/question numbers. Please ensure the steps to be idempotent — i.e., they are re-run safe and generate the same result if you run them again. If you get stuck on one question, move on to other questions and come back to this later (also there are aways of bypassing a step). Do not leave questions unanswered. Providing an answer, even if you cannot test it. Provide partial answers if you don’t know all of it. Each step counts! At the end of the exam, export your notebook as an HTML and upload it using the File Upload box of the mid-term exam assignment. A. Using Spark SQL to explore the Retail Dataset (20 points) This synthetic retail dataset is located at /databricks-datasets/retail-org/. It contains a collection of datasets (folders) representing different dimensions and facts for a retail organization. Among these: Sales Orders (under sales_orders): records the customers’ originating purchase orders. Products (under products): contains products that the company sells. Active Promotions (under active_promotions): shows how customers are progressing toward becoming eligible for promotions. As the first step, we want to inspect the data file format so that we know how to load it. 1. Set up the volume Set up a volume midterm under the schema spark of the catalog msbabigdata. 2. Display the files in /databricks-datasets/retail-org/products 3. Display the first 2000 bytes of products.csv This helps us determine how to load the csv file. 4. Create a new dataframe products using the file(s) in products Then, verify the dataframe by displaying the first 10 rows. 5. Save products as a Spark SQL table named products in the midterm database You need to create the midterm database first. After the table is created, verify it by selecting the first 10 rows from the table. Next, we will inspect and load active_promotions. 6. Display the files in /databricks-datasets/retail-org/active_promotions This helps you determine how to load this data source. 7. Create a new dataframe promotions using the file(s) in active_promotions Verify the dataframe by displaying the first 10 rows. 8. Join the two dataframes products and promotions Note that products.product_id corresponds to promotions.promo_item. Use the joined dataframes to calculate the promotion quantity for each product category (label: promo_quantity), and the total number of promotion sales line items for the category (label: num_entries). Order results by descending promo_quantity. 9. Save the joined dataframe as CSV files under the midterm volume, in the subfolder category_promotions. The CSV file needs to have a header row. You may verify it visually by opening it in the Unity Catalog. B. Using Spark MLLib to Predict Churn (20 points) In this task, we ask you to train a logisticRegression model with Spark’s MLlib (pyspark.ml) and test it on the test data. The data we use is a telecom customer churn dataset. It consists of customer activity data (features), along with a churn label specifying whether the customer canceled their wireless service subscription. We provide you with two datasets: the larger set (train.csv) for model training, and the smaller dataset (test.csv) for testing the model’s performance on unseen data. 10. Download data files and upload them to the midterm volume Download train.csv and test.csv from https://idsdl.csom.umn.edu/c/share/train.csv and https://idsdl.csom.umn.edu/c/share/test.csv, respectively. Upload them to the midterm volume in the Unity catalog. Verify the uploads by displaying the files in this volume. 11. Load train.csv and test.csv into data frames traindata and testdata, respectively. Then verify by displaying 10 rows from each DataFrame. 12. Create a predictive modeling pipeline that does the following: Use a feature transformer to rename columns — replace spaces with underscores. For example, Voice mail plan should be renamed to Voice_mail_plan If you use Spark SQL, you need to quote special column names with apostrophes (“`”) The last stage of the pipeline should be a crossvalidator that: tunes a logistic regression model that predicts churn All other columns should be used as features LogisticRegression can be found in pyspark.ml.classification explores a parameter grid that consists of three different values for the Logistic Regression’s regParam: 0.0, 0.01, 0.02. uses 3-fold cross-validation uses area under the ROC curve (under BinaryClassificationEvaluator) as the evaluation criterion. After building the pipeline, train it on the traindata to obtain a pipeline model. Although we do not require you to test each component of your pipeline, you may want to do that to make debugging easier, if there is any problem. You may not need to set up the caching directory for the crossvalidator, but just in case, please run the following. import os # Define your UC Volume path # Format: /Volumes//// uc_volume_path = “/Volumes/msbabigdata/spark/midterm” # Set the environment variable for SparkML caching os.environ[‘SPARKML_TEMP_DFS_PATH’] = uc_volume_path 13. Use the trained pipeline model to obtain predictions for the test dataset testdata. Then present the first 10 rows of the resulting DataFrame with the following fields: churn prediction 14. What is the area under the ROC curve for the test dataset? 15. Clean Up Drop all tables, databases, and volumes you have created in this exam.

You are seeing a patient with Amytrophic Lateral Scelorsis a…

You are seeing a patient with Amytrophic Lateral Scelorsis as an outpatient. The patient was referred for occupational therapy from their primary care doctor due to recent declines in fine motor skills and balance.  What approach and treatment selections are most appropriate for this level of care.