Default
Publications
- Ogden2023a - Layercake: Efficient Inference Serving with Cloud and Mobile Resources
- Ogden2021a - PieSlicer: Dynamically Improving Response Time for Cloud-based CNN Inference
- Ogden2021b - Many Models at the Edge: Scaling Deep Inference via Model-Level Caching
- Ogden2020 - MDInference: Balancing inference accuracy and latency for mobile applications
- Ogden2019 - CloudCoaster: Transient-aware Bursty Datacenter Workload Scheduling
- Ogden2019a - ModiPick: SLA-aware Accuracy Optimization For Mobile Deep Inference
- Ogden2019b - Characterizing the Deep Neural Networks Inference Performance of Mobile Applications
- Ogden2018 - MODI: Mobile Deep Inference Made Efficient by Edge Computing
- Tlachac2022a - Left on Read: Reply Latency for Anxiety & Depression Screening
- Tlachac2022a - Symptom Detection with Text Message Log Distributions for Holistic Depression and Anxiety Screening
- Gilman2020a - Demystifying the Placement Policies of the NVIDIA GPU Thread Block Scheduler for Concurrent Kernels
- Gilman2019 - Challenges and Opportunities of DNN Model Execution Caching
- Guo2019 - EdgeServe: efficient deep learning model caching at the edge