Main

related bits

0

processing priority

4

site type

0 (generic, awaiting analysis)

review version

11

html import

20 (imported)

Events

first seen date

2024-02-18 09:49:07

expired found date

-

created at

2024-07-02 12:30:54

updated at

2024-07-03 20:17:34

Domain name statistics

length

10

crc

56968

tld

2688

nm parts

0

nm random digits

0

nm rare letters

0

Connections

is subdomain of id

-

previous id

0

replaced with id

0

related id

-

dns primary id

168829982

dns alternative id

0

lifecycle status

0 (unclassified, or currently active)

Subdomains and pages

deleted subdomains

0

page imported products

0

page imported random

0

page imported parking

0

Error counters

count skipped due to recent timeouts on the same server IP

0

count content received but rejected due to 11-799

0

count dns errors

0

count cert errors

0

count timeouts

0

count http 429

0

count http 404

0

count http 403

0

count http 5xx

0

next operation date

-

Server

server bits

server ip

-

Mainpage statistics

mp import status

20

mp rejected date

-

mp saved date

-

mp size orig

13428

mp size raw text

2066

mp inner links count

5

mp inner links status

10 (links queued, awaiting import)

Open Graph

title

nocaps

description

Towards image captioning in the wild.

site name

nocaps

author

updated

2026-03-06 12:07:24

raw text

nocaps nocaps Overview Explore Download Paper Leaderboard Code The nocaps benchmark for novel object captioning (at scale). Image captioning models have achieved impressive results on datasets containing limited visual concepts and large amounts of paired image-caption training data. However, if these models are to ever function in the wild, a much larger variety of visual concepts must be learned, ideally from less supervision. To encourage the development of image captioning models that can learn visual concepts from alternative data sources, such as object detection datasets, we present the first large-scale benchmark for this task. Dubbed nocaps , for novel object captioning at scale, our benchmark consists of 166,100 human-generated captions describing 15,100 images from the Open Images validation and test sets. The associated training data consists of COCO image-caption pairs, plus Open Images image-level labels and object bounding boxes. ...

Text analysis

redirect type

0 (-)

block type

0 (no issues)

detected language

1 (English)

category id

Pozostałe (16)

index version

1

spam phrases

0

Text statistics

text nonlatin

0

text cyrillic

0

text characters

1588

text words

281

text unique words

159

text lines

57

text sentences

10

text paragraphs

2

text words per sentence

28

text matched phrases

0

text matched dictionaries

0

RSS

rss path

rss status

1 (priority 1 already searched, no matches found)

rss found date

-

rss size orig

0

rss items

0

rss spam phrases

0

rss detected language

0 (awaiting analysis)

inbefore feed id

-

inbefore status

0 (new)

Sitemap

sitemap path

sitemap status

1 (priority 1 already searched, no matches found)

sitemap review version

2

sitemap urls count

0

sitemap urls adult

0

sitemap filtered products

0

sitemap filtered videos

0

sitemap found date

-

sitemap process date

2024-07-02 12:30:54

sitemap first import date

-

sitemap last import date

-