aboutsummaryrefslogtreecommitdiffstats
path: root/main.go
diff options
context:
space:
mode:
authorAdam Mathes <adam@trenchant.org>2017-02-25 17:50:38 -0800
committerAdam Mathes <adam@trenchant.org>2017-02-25 17:50:38 -0800
commit8359198e5edb272311e10c77e6bf4cb662abd5ec (patch)
tree064589554fa36d5bf6c17436d39b68964034b1c2 /main.go
parent09af7d2f740eaa639fdcc2a7aa8c2ee1e3c3dd4d (diff)
downloadneko-8359198e5edb272311e10c77e6bf4cb662abd5ec.tar.gz
neko-8359198e5edb272311e10c77e6bf4cb662abd5ec.tar.bz2
neko-8359198e5edb272311e10c77e6bf4cb662abd5ec.zip
silent by default, verbose option added. -feeds cmd line option added
Diffstat (limited to 'main.go')
-rw-r--r--main.go35
1 files changed, 24 insertions, 11 deletions
diff --git a/main.go b/main.go
index 26dd1fc..d13efe7 100644
--- a/main.go
+++ b/main.go
@@ -7,36 +7,49 @@ import (
"adammathes.com/neko/models"
"adammathes.com/neko/models/feed"
"adammathes.com/neko/web"
- "log"
+ "adammathes.com/neko/vlog"
+ "fmt"
)
func main() {
- var serve, update bool
+ var serve, update, verbose, printFeeds bool
var configFile, newFeed string
flag.StringVar(&configFile, "c", "config.json", "`configuration` file")
flag.BoolVar(&update, "update", false, "update items by fetching feeds")
flag.BoolVar(&serve, "serve", false, "run http server")
-
+ flag.BoolVar(&verbose, "verbose", false, "verbose output")
+ flag.BoolVar(&printFeeds, "feeds", false, "list all currently crawled feeds")
flag.StringVar(&newFeed, "add", "", "add feed `http://example.com/rss.xml`")
flag.Parse()
- if !update && !serve && newFeed == ""{
+ // no command
+ if !update && !serve && !printFeeds && newFeed == ""{
flag.Usage()
return
}
config.Read(configFile)
models.InitDB(config.Config.DBServer)
-
+ vlog.VERBOSE = verbose
+
+ if update {
+ crawler.Crawl()
+ }
+ if serve {
+ vlog.Printf("starting web server at %s", config.Config.WebServer)
+ web.Serve()
+ }
if newFeed != "" {
feed.NewFeed(newFeed)
}
- if serve {
- log.Printf("starting web server at %s", config.Config.WebServer)
- web.Serve()
+ if printFeeds {
+ feeds,err := feed.All()
+ if err != nil {
+ panic(err)
+ }
+ for _,f := range(feeds) {
+ fmt.Printf("%s\n", f.Url)
+ }
}
- if update {
- crawler.Crawl()
- }
}