id
processing priority
3
site type
0 (generic, awaiting analysis)
review version
11
html import
20 (imported)
first seen date
2024-10-10 18:55:57
expired found date
-
created at
2024-10-10 18:55:57
updated at
2024-10-10 18:55:57
length
18
crc
26393
tld
86
nm parts
0
nm random digits
0
nm rare letters
0
is subdomain of id
87719371 (github.io)
previous id
0
replaced with id
0
related id
-
dns primary id
0
dns alternative id
0
lifecycle status
0 (unclassified, or currently active)
deleted subdomains
0
page imported products
0
page imported random
0
page imported parking
0
count skipped due to recent timeouts on the same server IP
0
count content received but rejected due to 11-799
0
count dns errors
0
count cert errors
0
count timeouts
0
count http 429
0
count http 404
0
count http 403
0
count http 5xx
0
next operation date
-
server bits
—
server ip
-
mp import status
20
mp rejected date
-
mp saved date
-
mp size orig
6293
mp size raw text
1459
mp inner links count
5
mp inner links status
10 (links queued, awaiting import)
title
Aim of the Workshop
description
Workshop at ICML 2022
image
site name
Hardware Aware Efficient Training (HAET)
author
updated
2026-03-05 11:11:37
raw text
Aim of the Workshop | Hardware Aware Efficient Training (HAET) Hardware Aware Efficient Training (HAET) Workshop at ICML 2022 Home CfP Keynotes Organizers Schedule Accepted Papers Aim of the Workshop To reach top-tier performance, deep learning models usually use a large number of parameters and operations, requiring considerable power and memory. Several works have proposed to tackle this problem using quantization of parameters, pruning, clustering of parameters, decompositions of convolutions, or distillation. However, most of these works aim at accelerating only the inference process and disregard the training phase. In practice, however, it is the learning phase that is by far the most complex. Despite recent efforts, promoting efficiency in the training process remains challenging. In this workshop, we propose to focus on reducing the complexity of the training process. Our aim is to gather researchers interested in reducing energy, time, or memory usage for...
redirect type
0 (-)
block type
0 (no issues)
detected language
1 (English)
category id
index version
1
spam phrases
0
text nonlatin
0
text cyrillic
0
text characters
1189
text words
210
text unique words
127
text lines
16
text sentences
11
text paragraphs
2
text words per sentence
19
text matched phrases
0
text matched dictionaries
0
links self subdomains
0
links other subdomains
0
links other domains
0
links spam adult
0
links spam random
0
links spam expired
0
links ext activities
0
links ext ecommerce
0
links ext finance
0
links ext crypto
0
links ext booking
0
links ext news
0
links ext leaks
0
links ext ugc
1
links ext klim
0
links ext generic
0
dol status
0
dol updated
2026-03-05 11:11:37
rss path
rss status
1 (priority 1 already searched, no matches found)
rss found date
-
rss size orig
0
rss items
0
rss spam phrases
0
rss detected language
0 (awaiting analysis)
inbefore feed id
-
inbefore status
0 (new)
sitemap path
sitemap status
1 (priority 1 already searched, no matches found)
sitemap review version
2
sitemap urls count
0
sitemap urls adult
0
sitemap filtered products
0
sitemap filtered videos
0
sitemap found date
-
sitemap process date
-
sitemap first import date
-
sitemap last import date
-