public inbox for gentoo-commits@lists.gentoo.org
 help / color / mirror / Atom feed
From: "Matt Turner" <mattst88@gentoo.org>
To: gentoo-commits@lists.gentoo.org
Subject: [gentoo-commits] proj/releng:master commit in: scripts/
Date: Sat, 31 Jul 2021 19:41:37 +0000 (UTC)	[thread overview]
Message-ID: <1627760476.36db6e7c8748e2d346e1ef92362807b0c242df76.mattst88@gentoo> (raw)

commit:     36db6e7c8748e2d346e1ef92362807b0c242df76
Author:     Matt Turner <mattst88 <AT> gentoo <DOT> org>
AuthorDate: Sat Jul 31 01:31:21 2021 +0000
Commit:     Matt Turner <mattst88 <AT> gentoo <DOT> org>
CommitDate: Sat Jul 31 19:41:16 2021 +0000
URL:        https://gitweb.gentoo.org/proj/releng.git/commit/?id=36db6e7c

scripts: Remove cruft

Presumably these were used once upon a time when catalyst builds were
done on infra-managed boxes (where puppet/chef made updating things
difficult).

Signed-off-by: Matt Turner <mattst88 <AT> gentoo.org>

 scripts/backup_snapshot_repo |  10 -
 scripts/cache-tools.py       | 700 -------------------------------------------
 scripts/run_catalyst         |   2 -
 scripts/run_official         |  39 ---
 scripts/run_snapshot         |   2 -
 scripts/stage_build.sh       | 162 ----------
 scripts/sudo_catalyst        |  28 --
 scripts/sudo_official        |  46 ---
 scripts/sudo_snapshot        |  20 --
 scripts/update_auto_tree     |  14 -
 scripts/update_official_tree |  14 -
 scripts/update_snapshot_tree |  14 -
 12 files changed, 1051 deletions(-)

diff --git a/scripts/backup_snapshot_repo b/scripts/backup_snapshot_repo
deleted file mode 100755
index 94b2828f..00000000
--- a/scripts/backup_snapshot_repo
+++ /dev/null
@@ -1,10 +0,0 @@
-#!/bin/bash
-
-# Start our rsyncs
-RSYNC_OPTS="--archive --delete --sparse --whole-file"
-
-if [ -e /release/repos/snapshot-tree/hooks/post-commit ]
-then
-	echo "$(date): Starting rsync of trees from tmpfs to disk..." >> /var/log/snapshot-tree-backup.log
-	rsync ${RSYNC_OPTS} /release/repos/snapshot-tree/* /release/repos/snapshot-tree-disk 2>&1 >> /var/log/snapshot-tree-backup.log || echo "$(date): rsync failed!" >> /var/log/snapshot-tree-backup.log
-fi

diff --git a/scripts/cache-tools.py b/scripts/cache-tools.py
deleted file mode 100755
index 03644503..00000000
--- a/scripts/cache-tools.py
+++ /dev/null
@@ -1,700 +0,0 @@
-#!/usr/bin/env python
-# Copyright 1999-2006 Gentoo Foundation
-# Distributed under the terms of the GNU General Public License v2
-# $Header: $
-#
-# Zac Medico <zmedico@gentoo.org>
-#
-
-import errno, fpformat, os, sys, time
-
-if not hasattr(__builtins__, "set"):
-	from sets import Set as set
-from itertools import chain
-
-def create_syncronized_func(myfunc, mylock):
-	def newfunc(*pargs, **kwargs):
-		mylock.acquire()
-		try:
-			myfunc(*pargs, **kwargs)
-		finally:
-			mylock.release()
-	return myfunc
-
-class ConsoleUpdate(object):
-
-	_synchronized_methods = ["append", "carriageReturn",
-		"newLine", "reset", "update"]
-
-	def __init__(self):
-		self.offset = 0
-		import sys
-		self.stream = sys.stdout
-		self.quiet = False
-		import threading
-		self._lock = threading.RLock()
-		for method_name in self._synchronized_methods:
-			myfunc = create_syncronized_func(
-				getattr(self, method_name), self._lock)
-			setattr(self, method_name, myfunc)
-		# ANSI code that clears from the cursor to the end of the line
-		self._CLEAR_EOL = None
-		try:
-			import curses
-			try:
-				curses.setupterm()
-				self._CLEAR_EOL = curses.tigetstr('el')
-			except curses.error:
-				pass
-		except ImportError:
-			pass
-		if not self._CLEAR_EOL:
-			self._CLEAR_EOL = '\x1b[K'
-
-	def acquire(self, **kwargs):
-		return self._lock.acquire(**kwargs)
-
-	def release(self):
-		self._lock.release()
-
-	def reset(self):
-		self.offset = 0
-
-	def carriageReturn(self):
-		if not self.quiet:
-			self.stream.write("\r")
-			self.stream.write(self._CLEAR_EOL)
-			self.offset = 0
-
-	def newLine(self):
-		if not self.quiet:
-			self.stream.write("\n")
-			self.stream.flush()
-			self.reset()
-
-	def update(self, msg):
-		if not self.quiet:
-			self.carriageReturn()
-			self.append(msg)
-
-	def append(self, msg):
-		if not self.quiet:
-			self.offset += len(msg)
-			self.stream.write(msg)
-			self.stream.flush()
-
-class ProgressCounter(object):
-	def __init__(self):
-		self.total = 0
-		self.current = 0
-
-class ProgressAnalyzer(ProgressCounter):
-	def __init__(self):
-		self.start_time = time.time()
-		self.currentTime = self.start_time
-		self._samples = []
-		self.sampleCount = 20
-	def percentage(self, digs=0):
-		if self.total > 0:
-			float_percent = 100 * float(self.current) / float(self.total)
-		else:
-			float_percent = 0.0
-		return fpformat.fix(float_percent,digs)
-	def totalTime(self):
-		self._samples.append((self.currentTime, self.current))
-		while len(self._samples) > self.sampleCount:
-			self._samples.pop(0)
-		prev_time, prev_count = self._samples[0]
-		time_delta = self.currentTime - prev_time
-		if time_delta > 0:
-			rate = (self.current - prev_count) / time_delta
-			if rate > 0:
-				return self.total / rate
-		return 0
-	def remaining_time(self):
-		return self.totalTime() - self.elapsed_time()
-	def elapsed_time(self):
-		return self.currentTime - self.start_time
-
-class ConsoleProgress(object):
-	def __init__(self, name="Progress", console=None):
-		self.name = name
-		self.analyzer = ProgressAnalyzer()
-		if console is None:
-			self.console = ConsoleUpdate()
-		else:
-			self.console = console
-		self.time_format="%H:%M:%S"
-		self.quiet = False
-		self.lastUpdate = 0
-		self.latency = 0.5
-
-	def formatTime(self, t):
-		return time.strftime(self.time_format, time.gmtime(t))
-
-	def displayProgress(self, current, total):
-		if self.quiet:
-			return
-
-		self.analyzer.currentTime = time.time()
-		if self.analyzer.currentTime - self.lastUpdate < self.latency:
-			return
-		self.lastUpdate = self.analyzer.currentTime
-		self.analyzer.current = current
-		self.analyzer.total = total
-
-		output = ((self.name, self.analyzer.percentage(1).rjust(4) + "%"),
-		("Elapsed", self.formatTime(self.analyzer.elapsed_time())),
-		("Remaining", self.formatTime(self.analyzer.remaining_time())),
-		("Total", self.formatTime(self.analyzer.totalTime())))
-		self.console.update(" ".join([ x[0] + ": " + x[1] for x in output ]))
-
-class ProgressHandler(object):
-	def __init__(self):
-		self.curval = 0
-		self.maxval = 0
-		self.last_update = 0
-		self.min_display_latency = 0.2
-
-	def onProgress(self, maxval, curval):
-		self.maxval = maxval
-		self.curval = curval
-		cur_time = time.time()
-		if cur_time - self.last_update >= self.min_display_latency:
-			self.last_update = cur_time
-			self.display()
-
-	def display(self):
-		raise NotImplementedError(self)
-
-def open_file(filename=None):
-	if filename is None:
-		f = sys.stderr
-	elif filename == "-":
-		f = sys.stdout
-	else:
-		try:
-			filename = os.path.expanduser(filename)
-			f = open(filename, "a")
-		except (IOError, OSError), e:
-			sys.stderr.write("%s\n" % e)
-			sys.exit(e.errno)
-	return f
-
-def create_log(name="", logfile=None, loglevel=0):
-	import logging
-	log = logging.getLogger(name)
-	log.setLevel(loglevel)
-	handler = logging.StreamHandler(open_file(logfile))
-	handler.setFormatter(logging.Formatter("%(levelname)s %(message)s"))
-	log.addHandler(handler)
-	return log
-
-def is_interrupt(e):
-	if isinstance(e, (SystemExit, KeyboardInterrupt)):
-		return True
-	return hasattr(e, "errno") and e.errno == errno.EINTR
-
-def mirror_cache(valid_nodes_iterable, src_cache, trg_cache, log,
-	eclass_cache, cleanse_on_transfer_failure):
-
-	cleanse_candidates = set(trg_cache.iterkeys())
-	update_count = 0
-
-	# Since the loop below is mission critical, we continue after *any*
-	# exception that is not an interrupt.
-
-	for x in valid_nodes_iterable:
-		log.debug("%s mirroring" % x)
-		if not cleanse_on_transfer_failure:
-			cleanse_candidates.discard(x)
-
-		try:
-			entry = copy_dict(src_cache[x])
-		except KeyError, e:
-			log.error("%s missing source: %s" % (x, str(e)))
-			del e
-			continue
-		except Exception, e:
-			if is_interrupt(e):
-				raise
-			log.error("%s reading source: %s" % (x, str(e)))
-			del e
-			continue
-
-		write_it = True
-		trg = None
-
-		try:
-			trg = copy_dict(trg_cache[x])
-			if long(trg["_mtime_"]) == long(entry["_mtime_"]) and \
-				eclass_cache.is_eclass_data_valid(trg["_eclasses_"]) and \
-				set(trg["_eclasses_"]) == set(entry["_eclasses_"]):
-				write_it = False
-		except KeyError:
-			pass
-		except Exception, e:
-			if is_interrupt(e):
-				raise
-			log.error("%s reading target: %s" % (x, str(e)))
-			del e
-
-		if trg and not write_it:
-			""" We don't want to skip the write unless we're really sure that
-			the existing cache is identical, so don't trust _mtime_ and
-			_eclasses_ alone."""
-			for d in (entry, trg):
-				if "EAPI" in d and d["EAPI"] in ("", "0"):
-					del d["EAPI"]
-			for k in set(chain(entry, trg)).difference(
-				("_mtime_", "_eclasses_")):
-				if trg.get(k, "") != entry.get(k, ""):
-					write_it = True
-					break
-
-		if write_it:
-			update_count += 1
-			log.info("%s transferring" % x)
-			inherited = entry.get("INHERITED", None)
-			if inherited:
-				if src_cache.complete_eclass_entries:
-					if not "_eclasses_" in entry:
-						log.error("%s missing _eclasses_" % x)
-						continue
-					if not eclass_cache.is_eclass_data_valid(entry["_eclasses_"]):
-						log.error("%s stale _eclasses_" % x)
-						continue
-				else:
-					entry["_eclasses_"] = eclass_cache.get_eclass_data(entry["INHERITED"].split(), \
-						from_master_only=True)
-					if not entry["_eclasses_"]:
-						log.error("%s stale _eclasses_" % x)
-						continue
-			try:
-				trg_cache[x] = entry
-				cleanse_candidates.discard(x)
-			except Exception, e:
-				if is_interrupt(e):
-					raise
-				log.error("%s writing target: %s" % (x, str(e)))
-				del e
-		else:
-			cleanse_candidates.discard(x)
-
-	if not trg_cache.autocommits:
-		try:
-			trg_cache.commit()
-		except Exception, e:
-			if is_interrupt(e):
-				raise
-			log.error("committing target: %s" % str(e))
-			del e
-
-	return update_count, cleanse_candidates
-
-def copy_dict(src, dest=None):
-	"""Some cache implementations throw cache errors when accessing the values.
-	We grab all the values at once here so that we don't have to be concerned
-	about exceptions later."""
-	if dest is None:
-		dest = {}
-	for k, v in src.iteritems():
-		dest[k] = v
-	return dest
-
-class ListPackages(object):
-	def __init__(self, portdb, log, shuffle=False):
-		self._portdb = portdb
-		self._log = log
-		self._shuffle = shuffle
-
-	def run(self):
-		log = self._log
-		cp_list = self._portdb.cp_list
-		cp_all = self._portdb.cp_all()
-		if self._shuffle:
-			from random import shuffle
-			shuffle(cp_all)
-		else:
-			cp_all.sort()
-		cpv_all = []
-		# Since the loop below is mission critical, we continue after *any*
-		# exception that is not an interrupt.
-		for cp in cp_all:
-			log.debug("%s cp_list" % cp)
-			try:
-				cpv_all.extend(cp_list(cp))
-			except Exception, e:
-				if is_interrupt(e):
-					raise
-				self._log.error("%s cp_list: %s" % (cp, str(e)))
-
-		self.cpv_all = cpv_all
-
-class MetadataGenerate(object):
-	"""When cache generation fails for some reason, cleanse the stale cache
-	entry if it exists.  This prevents the master mirror from distributing
-	stale cache, and will allow clients to safely assume that all cache is
-	valid.  The mtime requirement is especially annoying due to bug #139134
-	(timestamps of cache entries don't change when an eclass changes) and the
-	interaction of timestamps with rsync."""
-	def __init__(self, portdb, cpv_all, log):
-		self._portdb = portdb
-		self._cpv_all = cpv_all
-		self._log = log
-
-	def run(self, onProgress=None):
-		log = self._log
-		portdb = self._portdb
-		cpv_all = self._cpv_all
-		auxdb = portdb.auxdb[portdb.porttree_root]
-		cleanse_candidates = set(auxdb.iterkeys())
-
-		# Since the loop below is mission critical, we continue after *any*
-		# exception that is not an interrupt.
-		maxval = len(cpv_all)
-		curval = 0
-		if onProgress:
-			onProgress(maxval, curval)
-		while cpv_all:
-			cpv = cpv_all.pop(0)
-			log.debug("%s generating" % cpv)
-			try:
-				portdb.aux_get(cpv, ["EAPI"])
-				# Cleanse if the above doesn't succeed (prevent clients from
-				# receiving stale cache, and let them assume it is valid).
-				cleanse_candidates.discard(cpv)
-			except Exception, e:
-				if is_interrupt(e):
-					raise
-				log.error("%s generating: %s" % (cpv, str(e)))
-				del e
-			curval += 1
-			if onProgress:
-				onProgress(maxval, curval)
-
-		self.target_cache = auxdb
-		self.dead_nodes = cleanse_candidates
-
-class MetadataTransfer(object):
-	def __init__(self, portdb, cpv_all, forward, cleanse_on_transfer_failure,
-		log):
-		self._portdb = portdb
-		self._cpv_all = cpv_all
-		self._log = log
-		self._forward = forward
-		self._cleanse_on_transfer_failure = cleanse_on_transfer_failure
-
-	def run(self, onProgress=None):
-		log = self._log
-		portdb = self._portdb
-		cpv_all = self._cpv_all
-		aux_cache = portdb.auxdb[portdb.porttree_root]
-		import portage
-		auxdbkeys = portage.auxdbkeys[:]
-		metadbmodule = portdb.mysettings.load_best_module("portdbapi.metadbmodule")
-		portdir_cache = metadbmodule(portdb.porttree_root, "metadata/cache",
-			auxdbkeys)
-
-		maxval = len(cpv_all)
-		curval = 0
-		if onProgress:
-			onProgress(maxval, curval)
-		class pkg_iter(object):
-			def __init__(self, pkg_list, onProgress=None):
-				self.pkg_list = pkg_list
-				self.maxval = len(pkg_list)
-				self.curval = 0
-				self.onProgress = onProgress
-			def __iter__(self):
-				while self.pkg_list:
-					yield self.pkg_list.pop()
-					self.curval += 1
-					if self.onProgress:
-						self.onProgress(self.maxval, self.curval)
-
-		if self._forward:
-			src_cache = portdir_cache
-			trg_cache = aux_cache
-		else:
-			src_cache = aux_cache
-			trg_cache = portdir_cache
-
-		""" This encapsulates validation of eclass timestamps and also fills in
-		missing data (mtimes and/or paths) as necessary for the given cache
-		format."""
-		eclass_cache = portage.eclass_cache.cache(portdb.porttree_root)
-
-		if not trg_cache.autocommits:
-			trg_cache.sync(100)
-
-		self.target_cache = trg_cache
-		self.update_count, self.dead_nodes = mirror_cache(
-			pkg_iter(cpv_all, onProgress=onProgress),
-			src_cache, trg_cache, log, eclass_cache,
-			self._cleanse_on_transfer_failure)
-
-class CacheCleanse(object):
-	def __init__(self, auxdb, dead_nodes, log):
-		self._auxdb = auxdb
-		self._dead_nodes = dead_nodes
-		self._log = log
-	def run(self):
-		auxdb = self._auxdb
-		log = self._log
-		for cpv in self._dead_nodes:
-			try:
-				log.info("%s cleansing" % cpv)
-				del auxdb[cpv]
-			except Exception, e:
-				if is_interrupt(e):
-					raise
-				log.error("%s cleansing: %s" % (cpv, str(e)))
-				del e
-
-def import_portage():
-	try:
-		from portage import data as portage_data
-	except ImportError:
-		import portage_data
-	# If we're not already root or in the portage group, we make the gid of the
-	# current process become portage_gid.
-	if os.getgid() != 0 and portage_data.portage_gid not in os.getgroups():
-		portage_data.portage_gid = os.getgid()
-		portage_data.secpass = 1
-
-	os.environ["PORTAGE_LEGACY_GLOBALS"] = "false"
-	import portage
-	del os.environ["PORTAGE_LEGACY_GLOBALS"]
-	return portage
-
-def create_portdb(portdir=None, cachedir=None, config_root=None,
-	target_root=None, profile=None, **kwargs):
-
-	if cachedir is not None:
-		os.environ["PORTAGE_DEPCACHEDIR"] = cachedir
-	if config_root is None:
-		config_root = os.environ.get("PORTAGE_CONFIGROOT", "/")
-	if target_root is None:
-		target_root = os.environ.get("ROOT", "/")
-	if profile is None:
-		profile = ""
-
-	portage = import_portage()
-	try:
-		from portage import const as portage_const
-	except ImportError:
-		import portage_const
-
-	# Disable overlays because we only generate metadata for the main repo.
-	os.environ["PORTDIR_OVERLAY"] = ""
-	conf = portage.config(config_profile_path=profile,
-		config_incrementals=portage_const.INCREMENTALS,
-		target_root=target_root,
-		config_root=config_root)
-
-	if portdir is None:
-		portdir = conf["PORTDIR"]
-
-	# The cannonical path is the key for portdb.auxdb.
-	portdir = os.path.realpath(portdir)
-	conf["PORTDIR"] = portdir
-	conf.backup_changes("PORTDIR")
-
-	portdb = portage.portdbapi(portdir,
-		mysettings=conf)
-
-	return portdb
-
-def parse_args(myargv):
-	description = "This program will ensure that the metadata cache is up to date for entire portage tree."
-	usage = "usage: cache-tools [options] --generate || --transfer"
-	from optparse import OptionParser
-	parser = OptionParser(description=description, usage=usage)
-	parser.add_option("--portdir",
-		help="location of the portage tree",
-		dest="portdir")
-	parser.add_option("--cachedir",
-		help="location of the metadata cache",
-		dest="cachedir")
-	parser.add_option("--profile",
-		help="location of the profile",
-		dest="profile")
-	parser.add_option("--generate",
-		help="generate metadata as necessary to ensure that the cache is fully populated",
-		action="store_true", dest="generate", default=False)
-	parser.add_option("--shuffle",
-		help="generate cache in random rather than sorted order (useful to prevent two separate instances from competing to generate metadata for the same packages simultaneously)",
-		action="store_true", dest="shuffle", default=False)
-	parser.add_option("--transfer",
-		help="transfer metadata from portdir to cachedir or vice versa",
-		action="store_true", dest="transfer", default=False)
-	parser.add_option("--cleanse-on-transfer-failure",
-		help="cleanse target cache when transfer fails for any reason (such as the source being unavailable)",
-		action="store_true", dest="cleanse_on_transfer_failure", default=False)
-	parser.add_option("--forward",
-		help="forward metadata transfer flows from portdir to cachedir (default)",
-		action="store_true", dest="forward", default=True)
-	parser.add_option("--reverse",
-		help="reverse metadata transfer flows from cachedir to portdir",
-		action="store_false", dest="forward", default=True)
-	parser.add_option("--logfile",
-		help="send status messages to a file (default is stderr)",
-		dest="logfile", default=None)
-	parser.add_option("--loglevel",
-		help="numeric log level (defauls to 0 and may range from 0 to 50 corresponding to the default levels of the python logging module)",
-		dest="loglevel", default="0")
-	parser.add_option("--reportfile",
-		help="send a report to a file",
-		dest="reportfile", default=None)
-	parser.add_option("--spawn-outfile",
-		help="redirect ouput of spawned processes to a file instead of stdout/stderr",
-		dest="spawn_outfile", default=None)
-	parser.add_option("--no-progress",
-		action="store_false", dest="progress", default=True,
-		help="disable progress output to tty")
-	options, args = parser.parse_args(args=myargv)
-
-	# Conversion to dict allows us to use **opts as function args later on.
-	opts = {}
-	all_options = ("portdir", "cachedir", "profile", "progress", "logfile",
-		"loglevel", "generate", "transfer", "forward", "shuffle",
-		"spawn_outfile", "reportfile", "cleanse_on_transfer_failure")
-	for opt_name in all_options:
-		v = getattr(options, opt_name)
-		opts[opt_name] = v
-	return opts
-
-def run_command(args):
-	opts = parse_args(sys.argv[1:])
-
-	if opts["spawn_outfile"]:
-		fd = os.dup(1)
-		sys.stdout = os.fdopen(fd, 'w')
-		fd = os.dup(2)
-		sys.stderr = os.fdopen(fd, 'w')
-		f = open_file(opts["spawn_outfile"])
-		os.dup2(f.fileno(), 1)
-		os.dup2(f.fileno(), 2)
-		del fd, f
-
-	console = ConsoleUpdate()
-	if not opts["progress"] or not sys.stdout.isatty():
-		console.quiet = True
-	job = None
-	import signal, thread, threading
-	shutdown_initiated = threading.Event()
-	shutdown_complete = threading.Event()
-	def shutdown_console():
-		console.acquire()
-		try:
-			console.update("Interrupted.")
-			console.newLine()
-			console.quiet = True
-			shutdown_complete.set()
-			# Kill the main thread if necessary.
-			# This causes the SIGINT signal handler to be invoked in the
-			# main thread.  The signal handler needs to be an actual
-			# callable object (rather than something like signal.SIG_DFL)
-			# in order to avoid TypeError: 'int' object is not callable.
-			thread.interrupt_main()
-			thread.exit()
-		finally:
-			console.release()
-
-	def handle_interrupt(*args):
-		if shutdown_complete.isSet():
-			sys.exit(1)
-		# Lock the console from a new thread so that the main thread is allowed
-		# to cleanly complete any console interaction that may have been in
-		# progress when this interrupt arrived.
-		if not shutdown_initiated.isSet():
-			thread.start_new_thread(shutdown_console, ())
-			shutdown_initiated.set()
-
-	signal.signal(signal.SIGINT, handle_interrupt)
-	signal.signal(signal.SIGTERM, handle_interrupt)
-
-	try:
-		import datetime
-		datestamp = str(datetime.datetime.now())
-		time_begin = time.time()
-		log = create_log(name="MetadataGenerate",
-			logfile=opts["logfile"], loglevel=int(opts["loglevel"]))
-		if opts["reportfile"]:
-			reportfile = open_file(opts["reportfile"])
-		portdb = create_portdb(**opts)
-		try:
-			os.nice(int(portdb.mysettings.get("PORTAGE_NICENESS", "0")))
-		except (OSError, ValueError), e:
-			log.error("PORTAGE_NICENESS failed: '%s'" % str(e))
-			del e
-
-		job = ListPackages(portdb, log, shuffle=opts["shuffle"])
-		console.update("Listing packages in repository...")
-		job.run()
-		cpv_all = job.cpv_all
-		total_count = len(cpv_all)
-		if opts["generate"]:
-			job = MetadataGenerate(portdb, cpv_all, log)
-			name = "Cache generation"
-			complete_msg = "Metadata generation is complete."
-		elif opts["transfer"]:
-			job = MetadataTransfer(portdb, cpv_all, opts["forward"],
-			opts["cleanse_on_transfer_failure"], log)
-			if opts["forward"]:
-				name = "Forward transfer"
-				complete_msg = "Forward metadata transfer is complete."
-			else:
-				name = "Reverse transfer"
-				complete_msg = "Reverse metadata transfer is complete."
-		else:
-			sys.stderr.write("required options: --generate || --transfer\n")
-			sys.exit(os.EX_USAGE)
-		job.opts = opts
-
-		onProgress = None
-		if not console.quiet:
-			ui = ConsoleProgress(name=name, console=console)
-			progressHandler = ProgressHandler()
-			onProgress = progressHandler.onProgress
-			def display():
-				ui.displayProgress(progressHandler.curval, progressHandler.maxval)
-			progressHandler.display = display
-
-		job.run(onProgress=onProgress)
-
-		if not console.quiet:
-			# make sure the final progress is displayed
-			progressHandler.display()
-
-		update_count = None
-		if opts["transfer"]:
-			update_count = job.update_count
-		target_cache = job.target_cache
-		dead_nodes = job.dead_nodes
-		cleanse_count = len(dead_nodes)
-		console.update("Cleansing cache...")
-		job = CacheCleanse(target_cache, dead_nodes, log)
-		job.run()
-		console.update(complete_msg)
-		console.newLine()
-		time_end = time.time()
-		if opts["reportfile"]:
-			width = 20
-			reportfile.write(name.ljust(width) + "%s\n" % datestamp)
-			reportfile.write("Elapsed seconds".ljust(width) + "%f\n" % (time_end - time_begin))
-			reportfile.write("Total packages".ljust(width) + "%i\n" % total_count)
-			if update_count is not None:
-				reportfile.write("Updated packages".ljust(width) + "%i\n" % update_count)
-			reportfile.write("Cleansed packages".ljust(width) + "%i\n" % cleanse_count)
-			reportfile.write(("-"*50)+"\n")
-	except Exception, e:
-		if not is_interrupt(e):
-			raise
-		del e
-		handle_interrupt()
-	sys.exit(0)
-
-if __name__ == "__main__":
-	run_command(sys.argv[1:])

diff --git a/scripts/run_catalyst b/scripts/run_catalyst
deleted file mode 100755
index 997f6520..00000000
--- a/scripts/run_catalyst
+++ /dev/null
@@ -1,2 +0,0 @@
-#!/bin/bash
-sudo /release/bin/sudo_catalyst "$@"

diff --git a/scripts/run_official b/scripts/run_official
deleted file mode 100755
index dfb29f25..00000000
--- a/scripts/run_official
+++ /dev/null
@@ -1,39 +0,0 @@
-#!/bin/bash
-
-email_from="auto"
-email_to="releng@gentoo.org"
-url="https://poseidon.amd64.dev.gentoo.org/snapshots"
-snapshot_uri="/release/webroot/snapshots"
-svn_repo="/release/repos/snapshot-tree"
-
-send_email() {
-  subject="[Snapshot] ${1}"
-
-  echo -e "From: ${email_from}\r\nTo: ${email_to}\r\nSubject: ${subject}\r\n\r\nA new snapshot has been built from revision `svnlook history ${svn_repo} | head -n 3 | tail -n 1 | sed -e 's:^ *::' -e 's: .*$::'` of ${svn_repo}.  You can find it at ${url}.\r\n\r\n$(cat /release/snapshots/portage-${1}.tar.bz2.DIGESTS)\r\n" | /usr/sbin/sendmail -f ${email_from} ${email_to}
-}
-
-if [ "${email_from}" == "auto" ]
-then
-	username="$(whoami)"
-	if [ "${username}" == "root" ]
-	then
-		email_from="catalyst@poseidon.amd64.dev.gentoo.org"
-	else
-		email_from="${username}@gentoo.org"
-	fi
-fi
-
-sudo /release/bin/sudo_official "$@" && \
-echo "Starting rsync from /release/snapshots/portage-${1}.tar.bz2* to ${snapshot_uri}" && \
-rsync --archive --stats --progress /release/snapshots/portage-${1}.tar.bz2* \
-	${snapshot_uri}
-ret=$?
-
-if [ "${email_from}" == "none" ]
-then
-	echo "Skipping email step as configured..."
-else
-	[ $ret -eq 0 ] && send_email ${1}
-fi
-
-exit $ret

diff --git a/scripts/run_snapshot b/scripts/run_snapshot
deleted file mode 100755
index 20cc4609..00000000
--- a/scripts/run_snapshot
+++ /dev/null
@@ -1,2 +0,0 @@
-#!/bin/bash
-sudo /release/bin/sudo_snapshot "$@"

diff --git a/scripts/stage_build.sh b/scripts/stage_build.sh
deleted file mode 100755
index 46902c1e..00000000
--- a/scripts/stage_build.sh
+++ /dev/null
@@ -1,162 +0,0 @@
-#!/bin/bash
-
-PID=$$
-
-profile=
-version_stamp=
-subarch=
-stage1_seed=
-snapshot=
-config=/etc/catalyst/catalyst.conf
-email_from="catalyst@localhost"
-email_to="root@localhost"
-verbose=0
-
-usage() {
-  msg=$1
-
-  if [ -n "${msg}" ]; then
-    echo -e "${msg}\n";
-  fi
-
-  cat <<EOH
-Usage:
-  stage_build [-p|--profile <profile>] [-v|--version-stamp <stamp>]
-              [-a|--arch <arch>] [-s|--stage1-seed <seed>] [--verbose]
-              [-f|--email-from <from>] [-t|--email-to <to>] [-h|--help]
-
-Options:
-  -p|--profile        Sets the portage profile (required)
-  -v|--version-stamp  Sets the version stamp (required)
-  -a|--arch           Sets the 'subarch' in the spec (required)
-  -s|--stage1-seed    Sets the seed for the stage1 (required)
-  -S|--snapshot       Sets the snapshot name (if not given defaults to today's
-                      date)
-  -c|--config         catalyst config to use, defaults to catalyst default
-  --verbose           Send output of commands to console as well as log
-  -f|--email-from     Sets the 'From' on emails sent from this script (defaults
-                      to catalyst@localhost)
-  -t|--email-to       Sets the 'To' on emails sent from this script (defaults
-                      to root@localhost)
-  -h|--help           Show this message and quit
-
-Example:
-  stage_build -p default-linux/x86/2006.1 -v 2007.0_pre -a i686 -s default/stage3-i686-2006.1
-EOH
-}
-
-send_email() {
-  subject="[${subarch}] $1"
-  body=$2
-
-  echo -e "From: ${email_from}\r\nTo: ${email_to}\r\nSubject: ${subject}\r\n\r\nArch: ${subarch}\r\nProfile: ${profile}\r\nVersion stamp: ${version_stamp}\r\nStage1 seed: ${stage1_seed}\r\nSnapshot: ${snapshot}\r\n\r\n${body}\r\n" | /usr/sbin/sendmail -f ${email_from} ${email_to}
-}
-
-run_cmd() {
-  cmd=$1
-  logfile=$2
-
-  if [ $verbose = 1 ]; then
-    ${cmd} 2>&1 | tee ${logfile}
-  else
-    ${cmd} &> ${logfile}
-  fi
-}
-
-# Parse args
-params=${#}
-while [ ${#} -gt 0 ]
-do
-  a=${1}
-  shift
-  case "${a}" in
-    -h|--help)
-      usage
-      exit 0
-      ;;
-    -p|--profile)
-      profile=$1
-      shift
-      ;;
-    -v|--version-stamp)
-      version_stamp=$1
-      shift
-      ;;
-    -a|--arch)
-      subarch=$1
-      shift
-      ;;
-    -f|--email-from)
-      email_from=$1
-      shift
-      ;;
-    -t|--email-to)
-      email_to=$1
-      shift
-      ;;
-    -s|--stage1-seed)
-      stage1_seed=$1
-      shift
-      ;;
-    -S|--snapshot)
-      snapshot=$1
-      shift
-      ;;
-    -c|--config)
-      config=$1
-      shift
-      ;;
-    --verbose)
-      verbose=1
-      ;;
-    -*)
-      echo "You have specified an invalid option: ${a}"
-      usage
-      exit 1
-      ;;
-    esac
-done
-
-# Make sure all required values were specified
-if [ -z "${profile}" ]; then
-  usage "You must specify a profile."
-  exit 1
-fi
-if [ -z "${version_stamp}" ]; then
-  usage "You must specify a version stamp."
-  exit 1
-fi
-if [ -z "${subarch}" ]; then
-  usage "You must specify an arch."
-  exit 1
-fi
-if [ -z "${stage1_seed}" ]; then
-  usage "You must specify a stage1 seed."
-  exit 1
-fi
-cd /tmp
-
-if [ -z "${snapshot}" ]; then
-  snapshot=`date +%Y%m%d`
-  run_cmd "catalyst -c ${config} -s '${snapshot}'" "/tmp/catalyst_build_snapshot.${PID}.log"
-  if [ $? != 0 ]; then
-    send_email "Catalyst build error - snapshot" "$(</tmp/catalyst_build_snapshot.${PID}.log)"
-    exit 1
-  fi
-fi
-
-for i in 1 2 3; do
-  echo -e "subarch: ${subarch}\ntarget: stage${i}\nversion_stamp: ${version_stamp}\nrel_type: default\nprofile: ${profile}\nsnapshot: ${snapshot}" > stage${i}.spec
-  if [ ${i} = 1 ]; then
-    echo "source_subpath: ${stage1_seed}" >> stage${i}.spec
-  else 
-    echo "source_subpath: default/stage$(expr ${i} - 1)-${subarch}-${version_stamp}" >> stage${i}.spec
-  fi
-  run_cmd "catalyst -a -c ${config} -f stage${i}.spec" "/tmp/catalyst_build_stage${i}.${PID}.log"
-  if [ $? != 0 ]; then
-    send_email "Catalyst build error - stage${i}" "$(tail -n 200 /tmp/catalyst_build_stage${i}.${PID}.log)\r\n\r\nFull build log at /tmp/catalyst_build_stage${i}.${PID}.log"
-    exit 1
-  fi
-done
-
-send_email "Catalyst build success" "Everything finished successfully."

diff --git a/scripts/sudo_catalyst b/scripts/sudo_catalyst
deleted file mode 100755
index 19ecc90a..00000000
--- a/scripts/sudo_catalyst
+++ /dev/null
@@ -1,28 +0,0 @@
-#!/bin/bash
-
-usage() {
-	echo "Usage: $(basename ${0}) <arch> <target> <spec>"
-	echo "Where arch is either amd64 or x86, target is default, dev, hardened,"
-	echo "or uclibc, and spec is your spec file."
-	echo
-}
-
-if [ -z "$1" -o -z "$2" -o -z "$3" ]
-then
-	usage
-else
-	target="$(grep target ${3} | cut -d' ' -f2)"
-	/usr/bin/catalyst -c /etc/catalyst/${1}-${2}.conf -f ${3} ${4} ${5}
-#	&& \
-#	case ${target} in
-#		stage*|grp*|livecd-stage2)
-#			echo "Cleaning out ${target} temp files"
-#			rel_type="$(grep rel_type ${3} | cut -d' ' -f2)"
-#			subarch="$(grep subarch ${3} | cut -d' ' -f2)"
-#			version="$(grep version ${3} | cut -d' ' -f2)"
-#			storedir="$(grep storedir /etc/catalyst/${1}-${2}.conf | cut -d\" -f2)"
-#			echo "Removing ${storedir}/tmp/${rel_type}/${target}-${subarch}-${version}"
-#			rm -rf ${storedir}/tmp/${rel_type}/${target}-${subarch}-${version}
-#		;;
-#	esac
-fi

diff --git a/scripts/sudo_official b/scripts/sudo_official
deleted file mode 100755
index 80e7ca04..00000000
--- a/scripts/sudo_official
+++ /dev/null
@@ -1,46 +0,0 @@
-#!/bin/bash
-
-tree="/release/trees/snapshot-tree"
-portdir="${tree}/${1/_beta2/}/portage"
-cache_args="--portdir=${portdir} --cachedir=/release/tmp/depcache"
-
-usage() {
-	echo "Usage: $(basename ${0}) <version>"
-}
-
-if [ -z "${1}" ]
-then
-	usage
-else
-	cd ${tree}
-	echo "Clearing out old metadata cache"
-	rm -rf ${portdir}/metadata/cache
-	echo "Performing a svn up on ${tree}"
-	svn up || exit 1
-	mkdir -p ${portdir}/metadata/cache
-	echo "Recreating portage metadata cache"
-	cache-tools.py ${cache_args} --generate || exit 1
-	cache-tools.py ${cache_args} --transfer --reverse \
-		--cleanse-on-transfer-failure || exit 1
-	if [ ! -d ${portdir}/metadata/cache/sys-kernel ]
-	then
-		echo "Metadata update failed! Bailing out!"
-		exit 1
-	fi
-	catalyst -c /etc/catalyst/snapshot-official.conf -s ${1} \
-		-C portdir="${portdir}" || exit 1
-	for i in amd64 x86
-	do
-		for j in default dev hardened uclibc
-		do
-			cd /release/buildroot/${i}-${j}/snapshots
-			rm -f portage-official.tar.bz2 portage-${1}.tar.bz2*
-			ln -sf /release/snapshots/portage-${1}.tar.bz2 \
-				portage-official.tar.bz2
-			ln -sf /release/snapshots/portage-${1}.tar.bz2 \
-				portage-${1}.tar.bz2
-			ln -sf /release/snapshots/portage-${1}.tar.bz2.DIGESTS \
-				portage-${1}.tar.bz2.DIGESTS
-		done
-	done
-fi

diff --git a/scripts/sudo_snapshot b/scripts/sudo_snapshot
deleted file mode 100755
index 1ba64855..00000000
--- a/scripts/sudo_snapshot
+++ /dev/null
@@ -1,20 +0,0 @@
-#!/bin/bash
-usage() {
-	echo "Usage: $(basename ${0}) <version>"
-}
-if [ -z "${1}" ]
-then
-	usage
-else
-	catalyst -c /etc/catalyst/snapshot.conf -s ${1}
-	for i in amd64 x86
-	do
-		for j in default dev hardened uclibc
-		do
-			cd /release/buildroot/${i}-${j}/snapshots
-			rm -f portage-${1}.tar.bz2
-			ln -sf /release/snapshots/portage-${1}.tar.bz2 \
-				portage-${1}.tar.bz2
-		done
-	done
-fi

diff --git a/scripts/update_auto_tree b/scripts/update_auto_tree
deleted file mode 100755
index 029ef703..00000000
--- a/scripts/update_auto_tree
+++ /dev/null
@@ -1,14 +0,0 @@
-#!/bin/bash
-PORTDIR="/release/trees/portage-auto/"
-TMPREPO=$(mktemp)
-trap "rm -f $TMPREPO" EXIT
-cat >$TMPREPO <<EOF
-[DEFAULT]
-main-repo = gentoo
-
-[gentoo]
-location = $PORTDIR
-sync-type = rsync
-sync-uri = rsync://rsync.gentoo.org/gentoo-portage
-EOF
-PORTAGE_REPOSITORIES="$(cat $TMPREPO)" FEATURES="$FEATURES -news" emerge --sync -q

diff --git a/scripts/update_official_tree b/scripts/update_official_tree
deleted file mode 100755
index 55ecea0b..00000000
--- a/scripts/update_official_tree
+++ /dev/null
@@ -1,14 +0,0 @@
-#!/bin/bash
-PORTDIR="/release/trees/portage-official/"
-TMPREPO=$(mktemp)
-trap "rm -f $TMPREPO" EXIT
-cat >$TMPREPO <<EOF
-[DEFAULT]
-main-repo = gentoo
-
-[gentoo]
-location = $PORTDIR
-sync-type = rsync
-sync-uri = rsync://rsync.gentoo.org/gentoo-portage
-EOF
-PORTAGE_REPOSITORIES="$(cat $TMPREPO)" FEATURES="$FEATURES -news" emerge --sync -q

diff --git a/scripts/update_snapshot_tree b/scripts/update_snapshot_tree
deleted file mode 100755
index 76cd00ea..00000000
--- a/scripts/update_snapshot_tree
+++ /dev/null
@@ -1,14 +0,0 @@
-#!/bin/bash
-PORTDIR="/release/trees/portage-snapshot/"
-TMPREPO=$(mktemp)
-trap "rm -f $TMPREPO" EXIT
-cat >$TMPREPO <<EOF
-[DEFAULT]
-main-repo = gentoo
-
-[gentoo]
-location = $PORTDIR
-sync-type = rsync
-sync-uri = rsync://rsync.gentoo.org/gentoo-portage
-EOF
-PORTAGE_REPOSITORIES="$(cat $TMPREPO)" FEATURES="$FEATURES -news" emerge --sync -q


             reply	other threads:[~2021-07-31 19:41 UTC|newest]

Thread overview: 63+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-07-31 19:41 Matt Turner [this message]
  -- strict thread matches above, loose matches on Subject: below --
2023-10-09  0:08 [gentoo-commits] proj/releng:master commit in: scripts/ Robin H. Johnson
2023-10-09  0:05 Robin H. Johnson
2023-10-09  0:04 Robin H. Johnson
2023-10-08 23:35 Robin H. Johnson
2023-10-08 23:32 Robin H. Johnson
2023-10-08 23:20 Robin H. Johnson
2023-10-08 23:17 Robin H. Johnson
2023-10-08 23:17 Robin H. Johnson
2023-10-08 23:10 Robin H. Johnson
2023-10-08 23:09 Robin H. Johnson
2023-10-08 23:07 Robin H. Johnson
2023-10-08 23:07 Robin H. Johnson
2022-10-13 15:52 Ben Kohler
2022-03-01 10:23 Andreas K. Hüttel
2021-11-12 18:42 Robin H. Johnson
2021-11-12 18:31 Andreas K. Hüttel
2021-11-12 18:31 Andreas K. Hüttel
2021-11-12 18:19 Andreas K. Hüttel
2021-11-12 18:00 Robin H. Johnson
2021-11-12 17:45 Andreas K. Hüttel
2021-11-12  3:32 Matt Turner
2021-11-12  0:15 Robin H. Johnson
2021-11-01 20:46 Andreas K. Hüttel
2021-10-31 17:45 Andreas K. Hüttel
2021-09-02 21:51 Andreas K. Hüttel
2021-07-31 16:05 Andreas K. Hüttel
2021-07-31 16:00 Andreas K. Hüttel
2021-05-09 21:24 Andreas K. Hüttel
2021-01-23 14:54 Andreas K. Hüttel
2020-12-19 20:18 Anthony G. Basile
2020-08-14 19:21 Matt Turner
2018-11-21 19:13 Matt Turner
2018-11-21 19:13 Matt Turner
2018-11-21 19:13 Matt Turner
2016-05-27 21:01 Robin H. Johnson
2016-05-27 20:56 Robin H. Johnson
2016-05-06 18:32 Robin H. Johnson
2016-03-20 23:53 Mike Frysinger
2016-03-20 23:53 Mike Frysinger
2016-03-20 23:53 Mike Frysinger
2016-03-20 23:53 Mike Frysinger
2016-03-20 23:53 Mike Frysinger
2016-03-20 22:42 Mike Frysinger
2016-03-20 22:40 Mike Frysinger
2016-03-20 22:39 Mike Frysinger
2016-03-20 22:39 Mike Frysinger
2015-10-22 10:00 Jorge Manuel B. S. Vicetto
2015-10-20 17:30 Robin H. Johnson
2015-10-02 22:45 Jorge Manuel B. S. Vicetto
2015-02-06 20:13 Jorge Manuel B. S. Vicetto
2015-02-06 17:33 Jorge Manuel B. S. Vicetto
2014-03-27  3:09 Jorge Manuel B. S. Vicetto
2013-08-17 19:13 Raúl Porcel
2013-08-17 18:59 Raúl Porcel
2013-08-17 16:52 Raúl Porcel
2013-08-05 12:02 Raúl Porcel
2013-08-04 18:01 Raúl Porcel
2013-04-18  8:33 Jorge Manuel B. S. Vicetto
2013-01-20  7:44 Raúl Porcel
2012-11-04 13:24 Jorge Manuel B. S. Vicetto
2012-07-22 17:36 Raúl Porcel
2012-07-09 17:49 Jorge Manuel B. S. Vicetto

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1627760476.36db6e7c8748e2d346e1ef92362807b0c242df76.mattst88@gentoo \
    --to=mattst88@gentoo.org \
    --cc=gentoo-commits@lists.gentoo.org \
    --cc=gentoo-dev@lists.gentoo.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox