至少5家数藏平台发行徐悲鸿数字藏品,谁是真谁是假?

财联社Published on 2022-06-02Last updated on 2022-06-02

Abstract

徐悲鸿的墨宝,谁都可以用来发行数字藏品?

近日,“徐悲鸿美术馆”在社交媒体发布版权声明称,存在某些数字平台以徐悲鸿的名义为噱头发售相关数字藏品,不仅损害了消费者的利益,同时也侵犯了徐悲鸿的名誉身份权及徐悲鸿后人依法取得的各项知识产权权利。 
随后,有网友在评论区询问“幻核发售的徐悲鸿先生的作品是真是假”,“徐悲鸿美术馆”则回复其并未授权“幻核”发售徐悲鸿先生的数字藏品。 
此回应引起了小范围的争论。究竟是幻核确有侵权,还是一场乌龙?数字藏品的版权又该如何确定? 
当事三方的回应 
幻核发行的徐悲鸿数字墨马藏品于5月30日15点开售,该系列藏品共有8款,每款均限量发行3620份,每份售价为128元。 
目前该藏品已全部售罄。 
区块链日报记者就徐悲鸿数字墨马藏品的授权问题询问幻核方面,官方回复称,幻核一直坚持合规合法原则,每期发售都检查过合作方授权链路的,“本次徐悲鸿藏品也是如此”。 
“我们发布的不是徐悲鸿美术馆的作品。”幻核App上的人工客服也同样表示,由于徐悲鸿先生过世已超过50周年,所以拍卖所得的拥有者具有独立授权来跟幻核合作的权利。该客服告诉记者,本次藏品“由发行方北京皇城艺术品交易中心授权”。 
对此,北京皇城艺术品交易中心的工作人员向记者确认,中心确实有向幻核授权(徐悲鸿数字墨马藏品的发售)。 
“由于目前徐悲鸿先生去世已经50年了,其财产性权利已经过了保护期,拥有徐悲鸿画作原件的所有人是有权利与各数字藏品平台进行合作的。”德恒北京办公室律师闫泽娟在接受记者采访时表示。 
根据《著作权法》第22条的规定,作者的署名权、修改权、保护作品完整权的保护期不受限制。《著作权法》第23条则规定,复制权、发行权、信息网络传播权、改编权等由著作权人享有的其他权利是有保护期的,如果是自然人个人作品,保护期为作者终生及其死亡后五十年。 
闫泽娟进一步解释道,若自然人的个人作品过了保护期后,任何人均可免费使用,但是使用过程中不得侵害作者的署名权、修改权、保护作品完整权,不得违反宪法和法律,不得损害公共利益。 
“在相关画作已进入公有领域的情况下,理论上任何人都可以对其拍照、扫描等数字化处理后用来发行藏品,并不需要获得徐悲鸿美术馆的额外授权,所以很多平台会自主发行相关作品。”分布科技CEO达鸿飞也表达了相同的看法。 
徐悲鸿画作的“火热” 
据区块链日报不完全统计,目前有包括商汤科技旗下的数字猫、鲸探、幻核、虚猕数藏、芒境·人民艺术馆等多个数字藏品平台在内都已发行过或正有徐悲鸿作品在发售。 
在达鸿飞看来,这是因为作品本身具有极佳观赏性和收藏价值,再加上国内数字藏品平台井喷,都寻求在平台上发布受欢迎的数字藏品。因此“徐悲鸿代表作奔马题材受到各个平台的青睐,争相上架”。 
鲸探在今年4月以“徐悲鸿美术馆”5幅馆藏原作为设计元素,发售了徐悲鸿画作系列数字藏品,分别为《逆风》、《风雨鸡鸣》、《侧目》、《鹰击长空》和《奔马》,价格均为18元,每款发售1万份。 
在“徐悲鸿美术馆”发布的版权声明微博的评论区,其证实了数字猫、鲸探上发行的徐悲鸿作品的数字藏品均有“徐悲鸿美术馆”运营方及艺术品合作方时代悲鸿(北京)文化艺术中心的授权。 
根据芒境官方,该平台于5月11日上线的“徐悲鸿·马”系列数字藏品的授权方是“时代悲鸿”。 
另外,虚猕官方表示徐悲鸿先生的《十二生肖》系列数字藏品将于6月1日起连续12天在虚猕平台上线,每天发售一款。同时还称,“此次发售由徐悲鸿长孙徐小阳先生独家授权”。 
值得注意的是,闫泽娟认为,由于徐悲鸿画作涉及的财产权五十年保护期已过,很难存在独家授权一说。“除非是对已过保护期的画作进行二次创作形成新作品,这种情况下与平台的合作有可能存在独家发授的情形”。

Related Reads

Where Is the AI Infrastructure Industry Chain Stuck?

The AI infrastructure (AI Infra) industry chain is facing unprecedented systemic bottlenecks, despite the rapid emergence of applications like DeepSeek and Seedance 2.0. The surge in global computing demand has exposed critical constraints across multiple layers of the supply chain—from core manufacturing equipment and data center cabling to specialty materials and cleanroom facilities. Key challenges include four major "walls": - **Memory Wall**: High-bandwidth memory (HBM) and DRAM face structural shortages as AI inference demand outpaces training, with new capacity not expected until 2027. - **Bandwidth Wall**: Data transfer speeds lag behind computing power, causing multi-level bottlenecks in-chip, between chips, and across data centers. - **Compute Wall**: Advanced chip manufacturing, reliant on EUV lithography and monopolized by ASML, remains the fundamental constraint, with supply chain fragility affecting production. - **Power Wall**: While energy demand from data centers is rising, power supply is a solvable near-term challenge through diversified energy infrastructure. Expansion is further hindered by shortages in testing equipment, IC substrates (critical for GPUs and seeing price hikes over 30%), specialty materials like low-CTE glass fiber, and high-end cleanroom facilities. Connection technologies are evolving, with copper cables resurging for short-range links due to cost and latency advantages, while optical solutions dominate long-range scenarios. Innovations like hollow-core fiber and advanced PCB technologies (e.g., glass substrates, mSAP) are emerging to meet bandwidth needs. In summary, AI Infra bottlenecks are multidimensional, spanning compute, memory, bandwidth, power, and supply chain logistics. Advanced chip manufacturing remains the core constraint, while substrate, material, and equipment shortages present immediate challenges. The industry is moving toward hybrid copper-optical solutions and accelerated domestic supply chain development.

marsbit30m ago

Where Is the AI Infrastructure Industry Chain Stuck?

marsbit30m ago

Autonomy or Compatibility: The Choice Facing China's AI Ecosystem Behind the Delay of DeepSeek V4

DeepSeek V4's repeated delay in early 2026 has sparked global discussions on "de-CUDA-ization" in AI. The highly anticipated trillion-parameter open-source model is undergoing deep adaptation to Huawei’s Ascend chips using the CANN framework, representing China’s first systematic attempt to run a core AI model outside the CUDA ecosystem. This shift, however, comes with significant engineering challenges. While the model uses a MoE architecture to reduce computational load, it places extreme demands on memory bandwidth, chip interconnects, and system scheduling—areas where NVIDIA’s mature CUDA ecosystem currently excels. Migrating to Ascend introduces complexities in hardware topology, communication latency, and software optimization due to CANN’s relative immaturity compared to CUDA. The move highlights a broader strategic dilemma: short-term compatibility with CUDA offers practical benefits and faster adoption, as seen in CANN’s efforts to emulate CUDA interfaces. Yet, long-term over-reliance on compatibility risks inheriting CUDA’s limitations and stifling native innovation. If global AI shifts away from transformer-based architectures, strict compatibility could lead to technological obsolescence. Despite these challenges, DeepSeek V4’s eventual release could demonstrate the viability of a full domestic AI stack and accelerate CANN’s ecosystem growth. However, true technological independence will require building an original software-hardware paradigm beyond compatibility—a critical task for China’s AI ambitions in the next 3-5 years.

marsbit48m ago

Autonomy or Compatibility: The Choice Facing China's AI Ecosystem Behind the Delay of DeepSeek V4

marsbit48m ago

Trading

Spot
Futures
活动图片