diff --git a/Procfile b/Procfile new file mode 100644 index 000000000..62e430aca --- /dev/null +++ b/Procfile @@ -0,0 +1 @@ +web: gunicorn 'app:create_app()' \ No newline at end of file diff --git a/app/__init__.py b/app/__init__.py index 2764c4cc8..d392f914d 100644 --- a/app/__init__.py +++ b/app/__init__.py @@ -1,8 +1,8 @@ from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_migrate import Migrate -import os from dotenv import load_dotenv +import os db = SQLAlchemy() @@ -30,5 +30,7 @@ def create_app(test_config=None): migrate.init_app(app, db) # Register Blueprints here - + from .routes import tasks_bp, goal_bp + app.register_blueprint(tasks_bp) + app.register_blueprint(goal_bp) return app diff --git a/app/models/goal.py b/app/models/goal.py index 8cad278f8..6abe60396 100644 --- a/app/models/goal.py +++ b/app/models/goal.py @@ -4,3 +4,5 @@ class Goal(db.Model): goal_id = db.Column(db.Integer, primary_key=True) + title = db.Column(db.String) + tasks = db.relationship("Task", backref="goal", lazy=True) diff --git a/app/models/task.py b/app/models/task.py index 39c89cd16..2b98e6001 100644 --- a/app/models/task.py +++ b/app/models/task.py @@ -3,4 +3,8 @@ class Task(db.Model): - task_id = db.Column(db.Integer, primary_key=True) + id = db.Column(db.Integer, primary_key=True, autoincrement=True) + title = db.Column(db.String) + description = db.Column(db.String) + completed_at = db.Column(db.DateTime, nullable=True, default=None) + goal_id = db.Column(db.Integer, db.ForeignKey("goal.goal_id"), nullable=True) diff --git a/app/routes.py b/app/routes.py index 8e9dfe684..ebca57b59 100644 --- a/app/routes.py +++ b/app/routes.py @@ -1,2 +1,224 @@ -from flask import Blueprint +from app import db +from app.models.task import Task +from app.models.goal import Goal +from flask import json, request, Blueprint, make_response, jsonify +from datetime import datetime +import os +import requests +tasks_bp = Blueprint("tasks_bp", __name__, url_prefix="/tasks") + +@tasks_bp.route("", methods=["GET", "POST"]) +def handle_tasks(): + if request.method == "GET": + title_from_url = request.args.get("title") + if title_from_url: + tasks = Task.query.filter_by(title=title_from_url) + else: + sort = request.args.get("sort") + if not sort: + tasks = Task.query.all() + elif sort == "asc": + tasks = Task.query.order_by(Task.title.asc()).all() + elif sort == "desc": + tasks = Task.query.order_by(Task.title.desc()).all() + else: + tasks = Task.query.all() + + tasks_response = [] + for task in tasks: + tasks_response.append({ + "id": task.id, + "title": task.title, + "description": task.description, + "is_complete": bool(task.completed_at) + }) + return jsonify(tasks_response) + + elif request.method == "POST": + request_body = request.get_json() + title = request_body.get("title") + description = request_body.get("description") + + if not title or not description or "completed_at" not in request_body: + return jsonify({"details": "Invalid data"}), 400 + + new_task = Task(title=title, + description=description, + completed_at=request_body["completed_at"]) + db.session.add(new_task) + db.session.commit() + commited_task = {"task": + {"id": new_task.id, + "title": new_task.title, + "description": new_task.description, + "is_complete": bool(new_task.completed_at) + }} + return jsonify(commited_task), 201 + +@tasks_bp.route("/", methods=["GET", "PUT", "DELETE"]) +def handle_task(tasks_id): + task = Task.query.get_or_404(tasks_id) + if request.method == "GET": + if task.goal_id != None: + selected_task = {"task": + {"id": task.id, + "goal_id": task.goal_id, + "title": task.title, + "description": task.description, + "is_complete": bool(task.completed_at) + }} + else: + selected_task = {"task": + {"id": task.id, + "title": task.title, + "description": task.description, + "is_complete": bool(task.completed_at) + }} + return jsonify(selected_task),200 + + elif request.method == "PUT": + request_body = request.get_json() + task.title = request_body["title"] + task.description = request_body["description"] + task.completed_at = request_body["completed_at"] + updated_task = {'task':{ + "id": task.id, + "title": task.title, + "description": task.description, + "is_complete": bool(task.completed_at) + }} + db.session.commit() + return jsonify(updated_task),200 + + elif request.method == "DELETE": + db.session.delete(task) + db.session.commit() + task_response_body = {"details": f'Task {task.id} "{task.title}" successfully deleted'} + return jsonify(task_response_body),200 + +@tasks_bp.route("//mark_complete", methods=["PATCH"]) +def handle_task_complete(task_id): + task = Task.query.get_or_404(task_id) + task.completed_at = datetime.now() + + db.session.commit() + + patched_task = {"task": { + "id": task.id, + "title": task.title, + "description": task.description, + "is_complete": True + }} + return jsonify(patched_task),200 + +@tasks_bp.route("//mark_incomplete", methods=["PATCH"]) +def handle_task_incomplete(task_id): + task = Task.query.get_or_404(task_id) + task.completed_at = None + + db.session.commit() + + patched_task = {"task": { + "id": task.id, + "title": task.title, + "description": task.description, + "is_complete": False + }} + return jsonify(patched_task),200 + +# Slack Portion +def post_to_slack(text): + slack_token = os.environ.get("SLACK_TOKEN_POST") + slack_path = "https://slack.com/api/chat.postMessage" + query_params = { + "channel": "task-notification", + "text": text, + } + headers = { + "Authorization": f"Bearer {slack_token}" + } + requests.post(slack_path, params=query_params, headers=headers) + +# Goals Route Portion +goal_bp = Blueprint("goal_bp", __name__, url_prefix="/goals") + +@goal_bp.route("", methods=["GET", "POST"]) +def handle_goals(): + if request.method == "GET": + goals = Goal.query.all() + goals_response = [] + for goal in goals: + goals_response.append({ + "id": goal.goal_id, + "title": goal.title, + }) + return jsonify(goals_response), 200 + elif request.method == "POST": + request_body = request.get_json() + title = request_body.get("title") + if not title: + return jsonify({"details": "Invalid data"}), 400 + new_goal = Goal(title=request_body["title"]) + + db.session.add(new_goal) + db.session.commit() + goal_response_body = {"goal": {"id": new_goal.goal_id, "title": new_goal.title}} + + return jsonify(goal_response_body), 201 + +@goal_bp.route("/", methods=["GET", "PUT", "DELETE"]) +def handle_goal(goal_id): + goal = Goal.query.get_or_404(goal_id) + if request.method == "GET": + selected_goal = {"goal": + {"title": goal.title, + "id": goal.goal_id + }} + return jsonify(selected_goal), 200 + elif request.method == "PUT": + request_body = request.get_json() + goal.title = request_body["title"] + updated_goal = {'goal':{ + "id": goal.goal_id, + "title": goal.title + }} + db.session.commit() + return jsonify(updated_goal),200 + + elif request.method == "DELETE": + db.session.delete(goal) + db.session.commit() + goal_response_body = {"details": f'Goal {goal.goal_id} "{goal.title}" successfully deleted'} + return jsonify(goal_response_body),200 + +@goal_bp.route("//tasks", methods=["GET", "POST"]) +def handle_goals_and_tasks(goal_id): + if request.method == "POST": + goal = Goal.query.get_or_404(goal_id) + request_body = request.get_json() + for id in request_body["task_ids"]: + task = Task.query.get(id) + goal.tasks.append(task) + db.session.add(goal) + db.session.commit() + + goal_task_response_body = {"id": goal.goal_id, "task_ids": request_body["task_ids"]} + return jsonify(goal_task_response_body), 200 + + elif request.method == "GET": + goal = Goal.query.get_or_404(goal_id) + tasks = goal.tasks + list_of_tasks = [] + + for task in tasks: + individual_task = { + "id": task.id, + "goal_id": goal.goal_id, + "title": task.title, + "description": task.description, + "is_complete": bool(task.completed_at) + } + list_of_tasks.append(individual_task) + goal_task_get_response_body = {"id": goal.goal_id, "title": goal.title,"tasks": list_of_tasks} + return jsonify(goal_task_get_response_body), 200 diff --git a/migrations/README b/migrations/README new file mode 100644 index 000000000..98e4f9c44 --- /dev/null +++ b/migrations/README @@ -0,0 +1 @@ +Generic single-database configuration. \ No newline at end of file diff --git a/migrations/alembic.ini b/migrations/alembic.ini new file mode 100644 index 000000000..f8ed4801f --- /dev/null +++ b/migrations/alembic.ini @@ -0,0 +1,45 @@ +# A generic, single database configuration. + +[alembic] +# template used to generate migration files +# file_template = %%(rev)s_%%(slug)s + +# set to 'true' to run the environment during +# the 'revision' command, regardless of autogenerate +# revision_environment = false + + +# Logging configuration +[loggers] +keys = root,sqlalchemy,alembic + +[handlers] +keys = console + +[formatters] +keys = generic + +[logger_root] +level = WARN +handlers = console +qualname = + +[logger_sqlalchemy] +level = WARN +handlers = +qualname = sqlalchemy.engine + +[logger_alembic] +level = INFO +handlers = +qualname = alembic + +[handler_console] +class = StreamHandler +args = (sys.stderr,) +level = NOTSET +formatter = generic + +[formatter_generic] +format = %(levelname)-5.5s [%(name)s] %(message)s +datefmt = %H:%M:%S diff --git a/migrations/env.py b/migrations/env.py new file mode 100644 index 000000000..8b3fb3353 --- /dev/null +++ b/migrations/env.py @@ -0,0 +1,96 @@ +from __future__ import with_statement + +import logging +from logging.config import fileConfig + +from sqlalchemy import engine_from_config +from sqlalchemy import pool +from flask import current_app + +from alembic import context + +# this is the Alembic Config object, which provides +# access to the values within the .ini file in use. +config = context.config + +# Interpret the config file for Python logging. +# This line sets up loggers basically. +fileConfig(config.config_file_name) +logger = logging.getLogger('alembic.env') + +# add your model's MetaData object here +# for 'autogenerate' support +# from myapp import mymodel +# target_metadata = mymodel.Base.metadata +config.set_main_option( + 'sqlalchemy.url', + str(current_app.extensions['migrate'].db.engine.url).replace('%', '%%')) +target_metadata = current_app.extensions['migrate'].db.metadata + +# other values from the config, defined by the needs of env.py, +# can be acquired: +# my_important_option = config.get_main_option("my_important_option") +# ... etc. + + +def run_migrations_offline(): + """Run migrations in 'offline' mode. + + This configures the context with just a URL + and not an Engine, though an Engine is acceptable + here as well. By skipping the Engine creation + we don't even need a DBAPI to be available. + + Calls to context.execute() here emit the given string to the + script output. + + """ + url = config.get_main_option("sqlalchemy.url") + context.configure( + url=url, target_metadata=target_metadata, literal_binds=True + ) + + with context.begin_transaction(): + context.run_migrations() + + +def run_migrations_online(): + """Run migrations in 'online' mode. + + In this scenario we need to create an Engine + and associate a connection with the context. + + """ + + # this callback is used to prevent an auto-migration from being generated + # when there are no changes to the schema + # reference: http://alembic.zzzcomputing.com/en/latest/cookbook.html + def process_revision_directives(context, revision, directives): + if getattr(config.cmd_opts, 'autogenerate', False): + script = directives[0] + if script.upgrade_ops.is_empty(): + directives[:] = [] + logger.info('No changes in schema detected.') + + connectable = engine_from_config( + config.get_section(config.config_ini_section), + prefix='sqlalchemy.', + poolclass=pool.NullPool, + ) + + with connectable.connect() as connection: + context.configure( + connection=connection, + target_metadata=target_metadata, + process_revision_directives=process_revision_directives, + **current_app.extensions['migrate'].configure_args + ) + + with context.begin_transaction(): + context.run_migrations() + + +if context.is_offline_mode(): + run_migrations_offline() +else: + run_migrations_online() diff --git a/migrations/script.py.mako b/migrations/script.py.mako new file mode 100644 index 000000000..2c0156303 --- /dev/null +++ b/migrations/script.py.mako @@ -0,0 +1,24 @@ +"""${message} + +Revision ID: ${up_revision} +Revises: ${down_revision | comma,n} +Create Date: ${create_date} + +""" +from alembic import op +import sqlalchemy as sa +${imports if imports else ""} + +# revision identifiers, used by Alembic. +revision = ${repr(up_revision)} +down_revision = ${repr(down_revision)} +branch_labels = ${repr(branch_labels)} +depends_on = ${repr(depends_on)} + + +def upgrade(): + ${upgrades if upgrades else "pass"} + + +def downgrade(): + ${downgrades if downgrades else "pass"} diff --git a/migrations/versions/b9bbd4042499_.py b/migrations/versions/b9bbd4042499_.py new file mode 100644 index 000000000..899a23352 --- /dev/null +++ b/migrations/versions/b9bbd4042499_.py @@ -0,0 +1,42 @@ +"""empty message + +Revision ID: b9bbd4042499 +Revises: +Create Date: 2021-06-14 19:58:08.425060 + +""" +from alembic import op +import sqlalchemy as sa + + +# revision identifiers, used by Alembic. +revision = 'b9bbd4042499' +down_revision = None +branch_labels = None +depends_on = None + + +def upgrade(): + # ### commands auto generated by Alembic - please adjust! ### + op.create_table('goal', + sa.Column('goal_id', sa.Integer(), nullable=False), + sa.Column('title', sa.String(), nullable=True), + sa.PrimaryKeyConstraint('goal_id') + ) + op.create_table('task', + sa.Column('id', sa.Integer(), autoincrement=True, nullable=False), + sa.Column('title', sa.String(), nullable=True), + sa.Column('description', sa.String(), nullable=True), + sa.Column('completed_at', sa.DateTime(), nullable=True), + sa.Column('goal_id', sa.Integer(), nullable=True), + sa.ForeignKeyConstraint(['goal_id'], ['goal.goal_id'], ), + sa.PrimaryKeyConstraint('id') + ) + # ### end Alembic commands ### + + +def downgrade(): + # ### commands auto generated by Alembic - please adjust! ### + op.drop_table('task') + op.drop_table('goal') + # ### end Alembic commands ###