I work for a corporation and my job is not data science related. My team has been using me as their data analysis person as I enjoy data. I have been figuring out data analysis on my own building data charts in Excel and some minor work in Power BI. I would like to get some data science abilities under my belt. Due to COVID my company has put a hold on any training at this time. I'm looking for what could be done at very low or no cost to get things going. I'm not exactly sure where I should start and what I should be focusing on as well. I do know my company uses Power BI and has databases that use SQL.
I think more than anything, I need a clear plan on what I need to learn to be competent and above average from a skill perspective. A list that says, do this, then that, then that, then that, done. LOL
Would appreciate any advice.
Touching on what Irnbru mentioned, the single biggest thing you can do which smaller companies in my experience regularly neglect is pulling information directly from the database to run automated reports. Knowing the data flow inside-out and becoming very comfortable with all of the relevant databases present in the organisation can have an absolutely monumental impact on what you do.
Becoming familiar with SQL (whatever variety is most relevant) is very easy to do for basic tasks which comprise most tasks (e.g. non-dynamic SQL procedures which, outside of a consulting company, is most likely what you'll be dealing with; getting into more automated code can get trickier).
Becoming familiar with Python (particularly Pandas for any basic functionality or reporting) is quick to learn but can have a monumental business impact for a company with little or poor automated processes.
Excel is very often used as a crutch, so where possible I would strongly, strongly recommend avoid using Excel where possible (saving into it? Fine, but doing anything with the actual file? Not so good). Unless you're building macros in Excel to do what is needed for you (personally, even though I like VBA, I found it a pain to learn initially with a high learning curve), it's often very easy to think "I've got most of this where it is, I'll just finish it off in Excel" which might save time the first time you do it but it means any time the same task needs to be done it's going to be time consuming and slower than just pressing run (and having to open multiple different items, waiting for tasks to finish, etc. is time consuming and frustrating); building it right the first time means it's done and can be set on a recurring schedule as needed.
If you're familiar with SQL and Python, one of the biggest things (even though it's a very basic idea and very simple) which transformed business reporting for the business was introducing a workflow of:
1. Developing an SQL process to extract as much (relevant) data as possible while minimising the hit on performance.
2. Using PYODBC or cx_Oracle to connect to the database and run the query (inserting any variables as needed).
3. Using Pandas to summarise all of the data as required (bonus: use your preferred library to create some visualisations in a sharable manner).
4. Dump the results in an Excel (or your preferred format).
5. Have a Macro which loops through and ''finishes' the files (e.g. visualisations, basic formatting, etc.; the format will always be broadly similar so Macros can be designed more easily to accommodate this).
Even getting to (4) brought an enormously enhanced data accessibility and data awareness within the department, as once you have something like that, it's very easy to set it up on a scheduler.