1
0
mirror of https://gitlab.com/Anson-Projects/zine.git synced 2025-07-27 16:51:26 +00:00

Compare commits

..

4 Commits

Author SHA1 Message Date
ee157336c0 New Blogs 2025-07-05 14:58:41 -06:00
74039edb30 Remove caching 2025-06-30 15:57:01 -06:00
6e0fd55e39 Let clippy fix 2025-06-30 15:55:41 -06:00
da63b300c6 Do not lint on scheduled jobs 2025-06-30 15:48:03 -06:00
5 changed files with 9 additions and 15 deletions

View File

@@ -11,11 +11,6 @@ variables:
DEFAULT_PIPELINE_NAME: "$CI_COMMIT_BRANCH - $CI_COMMIT_MESSAGE" DEFAULT_PIPELINE_NAME: "$CI_COMMIT_BRANCH - $CI_COMMIT_MESSAGE"
SCHEDULED_PIPELINE_NAME: "Daily scheduled build pipeline" SCHEDULED_PIPELINE_NAME: "Daily scheduled build pipeline"
cache:
paths:
- target/
- cargo/
build: build:
image: rust:latest image: rust:latest
stage: build stage: build
@@ -29,11 +24,7 @@ lint:
- rustup component add clippy - rustup component add clippy
- cargo clippy --all-targets -- -D warnings - cargo clippy --all-targets -- -D warnings
rules: rules:
- if: $SCHEDULED_BUILD_PIPELINE == 'true' - if: "$CI_COMMIT_BRANCH != $CI_DEFAULT_BRANCH"
allow_failure: true
- if: $SCHEDULED_BUILD_PIPELINE != 'true'
allow_failure: false
test: test:
image: rust:latest image: rust:latest

View File

@@ -112,6 +112,7 @@ https://rubenerd.com/feed/
https://saccade.com/blog/feed/ https://saccade.com/blog/feed/
https://samhenri.gold/feed.xml https://samhenri.gold/feed.xml
https://sd.ai/rss/ https://sd.ai/rss/
https://shield.ai/feed/
https://simonschreibt.de/feed/ https://simonschreibt.de/feed/
https://simonwillison.net/atom/entries/ https://simonwillison.net/atom/entries/
https://simplyexplained.com/atom.xml https://simplyexplained.com/atom.xml
@@ -139,11 +140,13 @@ https://wingolog.org/feed/atom
https://www.bitsaboutmoney.com/archive/rss/ https://www.bitsaboutmoney.com/archive/rss/
https://www.blogofholding.com/?feed=rss2 https://www.blogofholding.com/?feed=rss2
https://www.brendangregg.com/blog/rss.xml https://www.brendangregg.com/blog/rss.xml
https://www.construction-physics.com/feed
https://www.doscher.com/rss/ https://www.doscher.com/rss/
https://www.elidedbranches.com/feeds/posts/default https://www.elidedbranches.com/feeds/posts/default
https://www.evanjones.ca/index.rss https://www.evanjones.ca/index.rss
https://www.factorio.com/blog/rss https://www.factorio.com/blog/rss
https://www.hillelwayne.com/post/index.xml https://www.hillelwayne.com/post/index.xml
https://www.inkandswitch.com/index.xml
https://www.jeffgeerling.com/blog.xml https://www.jeffgeerling.com/blog.xml
https://www.joelonsoftware.com/feed/ https://www.joelonsoftware.com/feed/
https://www.jonashietala.se/feed.xml https://www.jonashietala.se/feed.xml
@@ -156,6 +159,7 @@ https://www.righto.com/feeds/posts/default
https://www.robinwils.com/rss.xml https://www.robinwils.com/rss.xml
https://www.scattered-thoughts.net/atom.xml https://www.scattered-thoughts.net/atom.xml
https://www.shubhro.com/feed.xml https://www.shubhro.com/feed.xml
https://www.thornewolf.com/rss/
https://www.wezm.net/v2/rss.xml https://www.wezm.net/v2/rss.xml
https://www.zachleat.com/web/feed/ https://www.zachleat.com/web/feed/
https://xeiaso.net/blog.rss https://xeiaso.net/blog.rss

View File

@@ -1 +0,0 @@
https://www.construction-physics.com/feed

View File

@@ -116,7 +116,7 @@ pub fn read_feed(path: &str) -> Vec<Post> {
.filter_map(|url| match web_fetchers::fetch_feed(url) { .filter_map(|url| match web_fetchers::fetch_feed(url) {
Ok(entries) => Some(entries), Ok(entries) => Some(entries),
Err(e) => { Err(e) => {
println!("Failed to fetch or parse feed {}: {}", url, e); println!("Failed to fetch or parse feed {url}: {e}");
None None
} }
}) })
@@ -131,7 +131,7 @@ pub fn read_feed(path: &str) -> Vec<Post> {
.par_iter() .par_iter()
.map(|entry| { .map(|entry| {
Post::from_entry(entry).map_err(|e| { Post::from_entry(entry).map_err(|e| {
log::warn!("Failed to process entry: {}", e); log::warn!("Failed to process entry: {e}");
e e
}) })
}) })

View File

@@ -18,7 +18,7 @@ fn test_if_feeds_are_in_alphabetical_order() {
urls.sort(); urls.sort();
for url in urls { for url in urls {
println!("{}", url); println!("{url}");
} }
panic!("feeds.txt was not sorted!") panic!("feeds.txt was not sorted!")
} }
@@ -50,7 +50,7 @@ fn test_if_duplicates_in_feeds_list() {
if !duplicates.is_empty() { if !duplicates.is_empty() {
for dupe in duplicates { for dupe in duplicates {
println!("{}", dupe) println!("{dupe}")
} }
panic!("Duplicate entries found!") panic!("Duplicate entries found!")
} }