An investigation of statistical learning curves: do we always need big data?

dc.contributor.authorLi, Yang
dc.date.accessioned2017-08-16T21:25:03Z
dc.date.available2017-08-16T21:25:03Z
dc.date.issued2017en
dc.description.abstractThe rapid revolutionary rapid Big Data technology has attracted increasing attention and widely been used in many industries. It is not only benefiting our life dramatically, but also posing new challenges to us at the same time. In many situations, dealing with these big and complex data can extremely difficult. However, do we really always need big data? This thesis attempted to investigate whether do we need a large dataset to build a model with acceptable accuracy, how the number of observations affect the performance of statistical predictive methods and use learning curves to describe this relationship. Some popular statis- tical learning methods were considered and applied on 3 large datasets. An efficient parallel coding strategy in R was also provided.en
dc.identifier.urihttp://hdl.handle.net/10092/13954
dc.identifier.urihttp://dx.doi.org/10.26021/3394
dc.languageEnglish
dc.language.isoen
dc.publisherUniversity of Canterburyen
dc.rightsAll Right Reserveden
dc.rights.urihttps://canterbury.libguides.com/rights/thesesen
dc.titleAn investigation of statistical learning curves: do we always need big data?en
dc.typeTheses / Dissertationsen
thesis.degree.disciplineStatisticsen
thesis.degree.grantorUniversity of Canterburyen
thesis.degree.levelMastersen
thesis.degree.nameMaster of Scienceen
uc.bibnumber2513125
uc.collegeFaculty of Engineeringen
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Li, Yang_Master's Thesis.pdf
Size:
745.32 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: