Using stock exchange data, train a model that predicts the stock price level in the future. You may use R language but don’t put in source code in the answers. Your source code will be submitted separately
a. Explore and visualize the dataset. Plot price level over time in training and testing datasets. Explain what steps you have taken in preprocessing and why.
b. Develop two different models using Long short-term memory (LSTM), and ARIMA algorithms. Explain your methodology in determining model architectures in detail. For instance, answer which parameters did you use in the models, how did you determine those values and how did you optimize them.
c. Visualize training and testing results by plotting actual vs predicted values as a function over time in test and train datasets for each model. Plot different models in different styles (dashes, lines, dots, etc.) or in different colors so that they can be distinguished
d. Using a table, compare average performances of the models using evaluation metrics of root mean square error (RMSE) and mean absolute error (MAE) for both testing and training datasets.
e. Discuss the model performances and your key findings. Interpret the meaning of results in your opinion
2. Using the transactions dataset, answer the following questions in the context of Market basket analysis. The dataset shows the transaction IDs and the list of products sold within the transactions, separated by a space. You may use R language but don’t put in source code in the answers. Your source code will be submitted separately.
a. Apply Apriori algorithm to the dataset and find all the frequent itemsets with the specified minimal support levels of (10%, 20%, 30%, 50%). Evaluate the results and discuss your findings.
b. Apply Eclat algorithm to the dataset and find all the frequent itemsets with the specified minimal support levels of (15%, 25%, 35%, 45%). Evaluate the results and discuss your findings.
c. Discuss the performance bottlenecks of Aprior/Eclat algorithms and explain how can they be improved? Describe potential issues of the algorithms which may occur on running large transaction datasets with millions of transactions and each transaction can contain tens or hundreds of items?
d. Association rules with high confidence are usually considered more interesting. That said, association rules with 100% confidence level may not be considered as interesting. Why do you think that is the case? What might be the reason? Discuss why association rules with 97% confidence level can be regarded more important?
e. [4 points]. Explain the concept of lift and how could you use the lift value in recommendation of books in e-commerce sites?
f. [3 points]. Also, research the reasons why Netflix created the House of Cards TV series starring Kevin Spacey and Robin Wright. How did the concept of lift factor into Netflix’s decision to make the show?
CS 340 Milestone One Guidelines and Rubric Overview: For this assignment, you will implement the fundamental operations of create, read, update,
Retail Transaction Programming Project Project Requirements: Develop a program to emulate a purchase transaction at a retail store. This
7COM1028 Secure Systems Programming Referral Coursework: Secure
Create a GUI program that:Accepts the following from a user:Item NameItem QuantityItem PriceAllows the user to create a file to store the sales receip
CS 340 Final Project Guidelines and Rubric Overview The final project will encompass developing a web service using a software stack and impleme